Commit 16ab3306 authored by Linus Torvalds's avatar Linus Torvalds

Merge master.kernel.org:/home/davem/BK/sparc-2.5

into penguin.transmeta.com:/home/penguin/torvalds/repositories/kernel/linux
parents f2ea1c6b 4c91e008
...@@ -18,7 +18,7 @@ export-objs := mca.o mtrr.o msr.o cpuid.o microcode.o i386_ksyms.o time.o ...@@ -18,7 +18,7 @@ export-objs := mca.o mtrr.o msr.o cpuid.o microcode.o i386_ksyms.o time.o
obj-y := process.o semaphore.o signal.o entry.o traps.o irq.o vm86.o \ obj-y := process.o semaphore.o signal.o entry.o traps.o irq.o vm86.o \
ptrace.o i8259.o ioport.o ldt.o setup.o time.o sys_i386.o \ ptrace.o i8259.o ioport.o ldt.o setup.o time.o sys_i386.o \
i386_ksyms.o i387.o bluesmoke.o dmi_scan.o \ pci-dma.o i386_ksyms.o i387.o bluesmoke.o dmi_scan.o \
bootflag.o bootflag.o
obj-$(CONFIG_MCA) += mca.o obj-$(CONFIG_MCA) += mca.o
......
O_TARGET := pci.o O_TARGET := pci.o
obj-y := dma.o i386.o obj-y := i386.o
ifdef CONFIG_VISWS ifdef CONFIG_VISWS
obj-y += visws.o obj-y += visws.o
......
...@@ -3273,6 +3273,368 @@ static void serverworks_agp_enable(u32 mode) ...@@ -3273,6 +3273,368 @@ static void serverworks_agp_enable(u32 mode)
* AGP devices and collect their data. * AGP devices and collect their data.
*/ */
#ifdef CONFIG_AGP_HP_ZX1
#ifndef log2
#define log2(x) ffz(~(x))
#endif
#define HP_ZX1_IOVA_BASE GB(1UL)
#define HP_ZX1_IOVA_SIZE GB(1UL)
#define HP_ZX1_GART_SIZE (HP_ZX1_IOVA_SIZE / 2)
#define HP_ZX1_SBA_IOMMU_COOKIE 0x0000badbadc0ffeeUL
#define HP_ZX1_PDIR_VALID_BIT 0x8000000000000000UL
#define HP_ZX1_IOVA_TO_PDIR(va) ((va - hp_private.iova_base) >> \
hp_private.io_tlb_shift)
static aper_size_info_fixed hp_zx1_sizes[] =
{
{0, 0, 0}, /* filled in by hp_zx1_fetch_size() */
};
static gatt_mask hp_zx1_masks[] =
{
{HP_ZX1_PDIR_VALID_BIT, 0}
};
static struct _hp_private {
struct pci_dev *ioc;
volatile u8 *registers;
u64 *io_pdir; // PDIR for entire IOVA
u64 *gatt; // PDIR just for GART (subset of above)
u64 gatt_entries;
u64 iova_base;
u64 gart_base;
u64 gart_size;
u64 io_pdir_size;
int io_pdir_owner; // do we own it, or share it with sba_iommu?
int io_page_size;
int io_tlb_shift;
int io_tlb_ps; // IOC ps config
int io_pages_per_kpage;
} hp_private;
static int __init hp_zx1_ioc_shared(void)
{
struct _hp_private *hp = &hp_private;
printk(KERN_INFO PFX "HP ZX1 IOC: IOPDIR shared with sba_iommu\n");
/*
* IOC already configured by sba_iommu module; just use
* its setup. We assume:
* - IOVA space is 1Gb in size
* - first 512Mb is IOMMU, second 512Mb is GART
*/
hp->io_tlb_ps = INREG64(hp->registers, HP_ZX1_TCNFG);
switch (hp->io_tlb_ps) {
case 0: hp->io_tlb_shift = 12; break;
case 1: hp->io_tlb_shift = 13; break;
case 2: hp->io_tlb_shift = 14; break;
case 3: hp->io_tlb_shift = 16; break;
default:
printk(KERN_ERR PFX "Invalid IOTLB page size "
"configuration 0x%x\n", hp->io_tlb_ps);
hp->gatt = 0;
hp->gatt_entries = 0;
return -ENODEV;
}
hp->io_page_size = 1 << hp->io_tlb_shift;
hp->io_pages_per_kpage = PAGE_SIZE / hp->io_page_size;
hp->iova_base = INREG64(hp->registers, HP_ZX1_IBASE) & ~0x1;
hp->gart_base = hp->iova_base + HP_ZX1_IOVA_SIZE - HP_ZX1_GART_SIZE;
hp->gart_size = HP_ZX1_GART_SIZE;
hp->gatt_entries = hp->gart_size / hp->io_page_size;
hp->io_pdir = phys_to_virt(INREG64(hp->registers, HP_ZX1_PDIR_BASE));
hp->gatt = &hp->io_pdir[HP_ZX1_IOVA_TO_PDIR(hp->gart_base)];
if (hp->gatt[0] != HP_ZX1_SBA_IOMMU_COOKIE) {
hp->gatt = 0;
hp->gatt_entries = 0;
printk(KERN_ERR PFX "No reserved IO PDIR entry found; "
"GART disabled\n");
return -ENODEV;
}
return 0;
}
static int __init hp_zx1_ioc_owner(u8 ioc_rev)
{
struct _hp_private *hp = &hp_private;
printk(KERN_INFO PFX "HP ZX1 IOC: IOPDIR dedicated to GART\n");
/*
* Select an IOV page size no larger than system page size.
*/
if (PAGE_SIZE >= KB(64)) {
hp->io_tlb_shift = 16;
hp->io_tlb_ps = 3;
} else if (PAGE_SIZE >= KB(16)) {
hp->io_tlb_shift = 14;
hp->io_tlb_ps = 2;
} else if (PAGE_SIZE >= KB(8)) {
hp->io_tlb_shift = 13;
hp->io_tlb_ps = 1;
} else {
hp->io_tlb_shift = 12;
hp->io_tlb_ps = 0;
}
hp->io_page_size = 1 << hp->io_tlb_shift;
hp->io_pages_per_kpage = PAGE_SIZE / hp->io_page_size;
hp->iova_base = HP_ZX1_IOVA_BASE;
hp->gart_size = HP_ZX1_GART_SIZE;
hp->gart_base = hp->iova_base + HP_ZX1_IOVA_SIZE - hp->gart_size;
hp->gatt_entries = hp->gart_size / hp->io_page_size;
hp->io_pdir_size = (HP_ZX1_IOVA_SIZE / hp->io_page_size) * sizeof(u64);
return 0;
}
static int __init hp_zx1_ioc_init(void)
{
struct _hp_private *hp = &hp_private;
struct pci_dev *ioc;
int i;
u8 ioc_rev;
ioc = pci_find_device(PCI_VENDOR_ID_HP, PCI_DEVICE_ID_HP_ZX1_IOC, NULL);
if (!ioc) {
printk(KERN_ERR PFX "Detected HP ZX1 AGP bridge but no IOC\n");
return -ENODEV;
}
hp->ioc = ioc;
pci_read_config_byte(ioc, PCI_REVISION_ID, &ioc_rev);
for (i = 0; i < PCI_NUM_RESOURCES; i++) {
if (pci_resource_flags(ioc, i) == IORESOURCE_MEM) {
hp->registers = (u8 *) ioremap(pci_resource_start(ioc,
i),
pci_resource_len(ioc, i));
break;
}
}
if (!hp->registers) {
printk(KERN_ERR PFX "Detected HP ZX1 AGP bridge but no CSRs\n");
return -ENODEV;
}
/*
* If the IOTLB is currently disabled, we can take it over.
* Otherwise, we have to share with sba_iommu.
*/
hp->io_pdir_owner = (INREG64(hp->registers, HP_ZX1_IBASE) & 0x1) == 0;
if (hp->io_pdir_owner)
return hp_zx1_ioc_owner(ioc_rev);
return hp_zx1_ioc_shared();
}
static int hp_zx1_fetch_size(void)
{
int size;
size = hp_private.gart_size / MB(1);
hp_zx1_sizes[0].size = size;
agp_bridge.current_size = (void *) &hp_zx1_sizes[0];
return size;
}
static int hp_zx1_configure(void)
{
struct _hp_private *hp = &hp_private;
agp_bridge.gart_bus_addr = hp->gart_base;
agp_bridge.capndx = pci_find_capability(agp_bridge.dev, PCI_CAP_ID_AGP);
pci_read_config_dword(agp_bridge.dev,
agp_bridge.capndx + PCI_AGP_STATUS, &agp_bridge.mode);
if (hp->io_pdir_owner) {
OUTREG64(hp->registers, HP_ZX1_PDIR_BASE,
virt_to_phys(hp->io_pdir));
OUTREG64(hp->registers, HP_ZX1_TCNFG, hp->io_tlb_ps);
OUTREG64(hp->registers, HP_ZX1_IMASK, ~(HP_ZX1_IOVA_SIZE - 1));
OUTREG64(hp->registers, HP_ZX1_IBASE, hp->iova_base | 0x1);
OUTREG64(hp->registers, HP_ZX1_PCOM,
hp->iova_base | log2(HP_ZX1_IOVA_SIZE));
INREG64(hp->registers, HP_ZX1_PCOM);
}
return 0;
}
static void hp_zx1_cleanup(void)
{
struct _hp_private *hp = &hp_private;
if (hp->io_pdir_owner)
OUTREG64(hp->registers, HP_ZX1_IBASE, 0);
iounmap((void *) hp->registers);
}
static void hp_zx1_tlbflush(agp_memory * mem)
{
struct _hp_private *hp = &hp_private;
OUTREG64(hp->registers, HP_ZX1_PCOM,
hp->gart_base | log2(hp->gart_size));
INREG64(hp->registers, HP_ZX1_PCOM);
}
static int hp_zx1_create_gatt_table(void)
{
struct _hp_private *hp = &hp_private;
int i;
if (hp->io_pdir_owner) {
hp->io_pdir = (u64 *) __get_free_pages(GFP_KERNEL,
get_order(hp->io_pdir_size));
if (!hp->io_pdir) {
printk(KERN_ERR PFX "Couldn't allocate contiguous "
"memory for I/O PDIR\n");
hp->gatt = 0;
hp->gatt_entries = 0;
return -ENOMEM;
}
memset(hp->io_pdir, 0, hp->io_pdir_size);
hp->gatt = &hp->io_pdir[HP_ZX1_IOVA_TO_PDIR(hp->gart_base)];
}
for (i = 0; i < hp->gatt_entries; i++) {
hp->gatt[i] = (unsigned long) agp_bridge.scratch_page;
}
return 0;
}
static int hp_zx1_free_gatt_table(void)
{
struct _hp_private *hp = &hp_private;
if (hp->io_pdir_owner)
free_pages((unsigned long) hp->io_pdir,
get_order(hp->io_pdir_size));
else
hp->gatt[0] = HP_ZX1_SBA_IOMMU_COOKIE;
return 0;
}
static int hp_zx1_insert_memory(agp_memory * mem, off_t pg_start, int type)
{
struct _hp_private *hp = &hp_private;
int i, k;
off_t j, io_pg_start;
int io_pg_count;
if (type != 0 || mem->type != 0) {
return -EINVAL;
}
io_pg_start = hp->io_pages_per_kpage * pg_start;
io_pg_count = hp->io_pages_per_kpage * mem->page_count;
if ((io_pg_start + io_pg_count) > hp->gatt_entries) {
return -EINVAL;
}
j = io_pg_start;
while (j < (io_pg_start + io_pg_count)) {
if (hp->gatt[j]) {
return -EBUSY;
}
j++;
}
if (mem->is_flushed == FALSE) {
CACHE_FLUSH();
mem->is_flushed = TRUE;
}
for (i = 0, j = io_pg_start; i < mem->page_count; i++) {
unsigned long paddr;
paddr = mem->memory[i];
for (k = 0;
k < hp->io_pages_per_kpage;
k++, j++, paddr += hp->io_page_size) {
hp->gatt[j] = agp_bridge.mask_memory(paddr, type);
}
}
agp_bridge.tlb_flush(mem);
return 0;
}
static int hp_zx1_remove_memory(agp_memory * mem, off_t pg_start, int type)
{
struct _hp_private *hp = &hp_private;
int i, io_pg_start, io_pg_count;
if (type != 0 || mem->type != 0) {
return -EINVAL;
}
io_pg_start = hp->io_pages_per_kpage * pg_start;
io_pg_count = hp->io_pages_per_kpage * mem->page_count;
for (i = io_pg_start; i < io_pg_count + io_pg_start; i++) {
hp->gatt[i] = agp_bridge.scratch_page;
}
agp_bridge.tlb_flush(mem);
return 0;
}
static unsigned long hp_zx1_mask_memory(unsigned long addr, int type)
{
return HP_ZX1_PDIR_VALID_BIT | addr;
}
static unsigned long hp_zx1_unmask_memory(unsigned long addr)
{
return addr & ~(HP_ZX1_PDIR_VALID_BIT);
}
static int __init hp_zx1_setup (struct pci_dev *pdev)
{
agp_bridge.masks = hp_zx1_masks;
agp_bridge.num_of_masks = 1;
agp_bridge.dev_private_data = NULL;
agp_bridge.size_type = FIXED_APER_SIZE;
agp_bridge.needs_scratch_page = FALSE;
agp_bridge.configure = hp_zx1_configure;
agp_bridge.fetch_size = hp_zx1_fetch_size;
agp_bridge.cleanup = hp_zx1_cleanup;
agp_bridge.tlb_flush = hp_zx1_tlbflush;
agp_bridge.mask_memory = hp_zx1_mask_memory;
agp_bridge.unmask_memory = hp_zx1_unmask_memory;
agp_bridge.agp_enable = agp_generic_agp_enable;
agp_bridge.cache_flush = global_cache_flush;
agp_bridge.create_gatt_table = hp_zx1_create_gatt_table;
agp_bridge.free_gatt_table = hp_zx1_free_gatt_table;
agp_bridge.insert_memory = hp_zx1_insert_memory;
agp_bridge.remove_memory = hp_zx1_remove_memory;
agp_bridge.alloc_by_type = agp_generic_alloc_by_type;
agp_bridge.free_by_type = agp_generic_free_by_type;
agp_bridge.agp_alloc_page = agp_generic_alloc_page;
agp_bridge.agp_destroy_page = agp_generic_destroy_page;
agp_bridge.cant_use_aperture = 1;
return hp_zx1_ioc_init();
(void) pdev; /* unused */
}
#endif /* CONFIG_AGP_HP_ZX1 */
pci_for_each_dev(device) { pci_for_each_dev(device) {
cap_ptr = pci_find_capability(device, PCI_CAP_ID_AGP); cap_ptr = pci_find_capability(device, PCI_CAP_ID_AGP);
...@@ -4087,6 +4449,15 @@ intel_850_setup }, ...@@ -4087,6 +4449,15 @@ intel_850_setup },
via_generic_setup }, via_generic_setup },
#endif /* CONFIG_AGP_VIA */ #endif /* CONFIG_AGP_VIA */
#ifdef CONFIG_AGP_HP_ZX1
{ PCI_DEVICE_ID_HP_ZX1_LBA,
PCI_VENDOR_ID_HP,
HP_ZX1,
"HP",
"ZX1",
hp_zx1_setup },
#endif
#ifdef CONFIG_AGP_HP_ZX1 #ifdef CONFIG_AGP_HP_ZX1
{ PCI_DEVICE_ID_HP_ZX1_LBA, { PCI_DEVICE_ID_HP_ZX1_LBA,
PCI_VENDOR_ID_HP, PCI_VENDOR_ID_HP,
...@@ -4320,6 +4691,23 @@ static int __init agp_find_supported_device(void) ...@@ -4320,6 +4691,23 @@ static int __init agp_find_supported_device(void)
#endif /* CONFIG_AGP_SWORKS */ #endif /* CONFIG_AGP_SWORKS */
#ifdef CONFIG_AGP_HP_ZX1
if (dev->vendor == PCI_VENDOR_ID_HP) {
do {
/* ZX1 LBAs can be either PCI or AGP bridges */
if (pci_find_capability(dev, PCI_CAP_ID_AGP)) {
printk(KERN_INFO PFX "Detected HP ZX1 AGP "
"chipset at %s\n", dev->slot_name);
agp_bridge.type = HP_ZX1;
agp_bridge.dev = dev;
return hp_zx1_setup(dev);
}
dev = pci_find_class(PCI_CLASS_BRIDGE_HOST << 8, dev);
} while (dev);
return -ENODEV;
}
#endif /* CONFIG_AGP_HP_ZX1 */
#ifdef CONFIG_AGP_HP_ZX1 #ifdef CONFIG_AGP_HP_ZX1
if (dev->vendor == PCI_VENDOR_ID_HP) { if (dev->vendor == PCI_VENDOR_ID_HP) {
do { do {
......
...@@ -317,6 +317,8 @@ drm_agp_head_t *DRM(agp_init)(void) ...@@ -317,6 +317,8 @@ drm_agp_head_t *DRM(agp_init)(void)
break; break;
#endif #endif
case HP_ZX1: head->chipset = "HP ZX1"; break;
default: head->chipset = "Unknown"; break; default: head->chipset = "Unknown"; break;
} }
#if LINUX_VERSION_CODE <= 0x020408 #if LINUX_VERSION_CODE <= 0x020408
......
...@@ -2692,7 +2692,7 @@ ace_load_tx_bd(struct ace_private *ap, struct tx_desc *desc, u64 addr, ...@@ -2692,7 +2692,7 @@ ace_load_tx_bd(struct ace_private *ap, struct tx_desc *desc, u64 addr,
flagsize &= ~BD_FLG_COAL_NOW; flagsize &= ~BD_FLG_COAL_NOW;
#endif #endif
if (!ACE_IS_TIGON_I(ap)) { if (ACE_IS_TIGON_I(ap)) {
writel(addr >> 32, &desc->addr.addrhi); writel(addr >> 32, &desc->addr.addrhi);
writel(addr & 0xffffffff, &desc->addr.addrlo); writel(addr & 0xffffffff, &desc->addr.addrlo);
writel(flagsize, &desc->flagsize); writel(flagsize, &desc->flagsize);
......
...@@ -29,7 +29,7 @@ obj-$(CONFIG_ALPHA) += setup-bus.o setup-irq.o ...@@ -29,7 +29,7 @@ obj-$(CONFIG_ALPHA) += setup-bus.o setup-irq.o
obj-$(CONFIG_ARM) += setup-bus.o setup-irq.o obj-$(CONFIG_ARM) += setup-bus.o setup-irq.o
obj-$(CONFIG_PARISC) += setup-bus.o obj-$(CONFIG_PARISC) += setup-bus.o
obj-$(CONFIG_SUPERH) += setup-bus.o setup-irq.o obj-$(CONFIG_SUPERH) += setup-bus.o setup-irq.o
obj-$(CONFIG_ALL_PPC) += setup-bus.o obj-$(CONFIG_PPC32) += setup-irq.o
obj-$(CONFIG_DDB5476) += setup-bus.o obj-$(CONFIG_DDB5476) += setup-bus.o
obj-$(CONFIG_SGI_IP27) += setup-irq.o obj-$(CONFIG_SGI_IP27) += setup-irq.o
......
...@@ -129,7 +129,7 @@ static void sd_rw_intr(Scsi_Cmnd * SCpnt); ...@@ -129,7 +129,7 @@ static void sd_rw_intr(Scsi_Cmnd * SCpnt);
static Scsi_Disk * sd_get_sdisk(int index); static Scsi_Disk * sd_get_sdisk(int index);
#if defined(CONFIG_PPC) #if defined(CONFIG_PPC32)
/** /**
* sd_find_target - find kdev_t of first scsi disk that matches * sd_find_target - find kdev_t of first scsi disk that matches
* given host and scsi_id. * given host and scsi_id.
...@@ -149,7 +149,7 @@ sd_find_target(void *hp, int scsi_id) ...@@ -149,7 +149,7 @@ sd_find_target(void *hp, int scsi_id)
{ {
Scsi_Disk *sdkp; Scsi_Disk *sdkp;
Scsi_Device *sdp; Scsi_Device *sdp;
Scsi_Host *shp = hp; struct Scsi_Host *shp = hp;
int dsk_nr; int dsk_nr;
unsigned long iflags; unsigned long iflags;
...@@ -162,7 +162,7 @@ sd_find_target(void *hp, int scsi_id) ...@@ -162,7 +162,7 @@ sd_find_target(void *hp, int scsi_id)
sdp = sdkp->device; sdp = sdkp->device;
if (sdp && (sdp->host == shp) && (sdp->id == scsi_id)) { if (sdp && (sdp->host == shp) && (sdp->id == scsi_id)) {
read_unlock_irqrestore(&sd_dsk_arr_lock, iflags); read_unlock_irqrestore(&sd_dsk_arr_lock, iflags);
return MKDEV_SD(k); return MKDEV_SD(dsk_nr);
} }
} }
read_unlock_irqrestore(&sd_dsk_arr_lock, iflags); read_unlock_irqrestore(&sd_dsk_arr_lock, iflags);
......
...@@ -38,60 +38,74 @@ static void redo_inode_mask(struct inode *inode) ...@@ -38,60 +38,74 @@ static void redo_inode_mask(struct inode *inode)
inode->i_dnotify_mask = new_mask; inode->i_dnotify_mask = new_mask;
} }
void dnotify_flush(struct file *filp, fl_owner_t id)
{
struct dnotify_struct *dn;
struct dnotify_struct **prev;
struct inode *inode;
inode = filp->f_dentry->d_inode;
if (!S_ISDIR(inode->i_mode))
return;
write_lock(&dn_lock);
prev = &inode->i_dnotify;
while ((dn = *prev) != NULL) {
if ((dn->dn_owner == id) && (dn->dn_filp == filp)) {
*prev = dn->dn_next;
redo_inode_mask(inode);
kmem_cache_free(dn_cache, dn);
break;
}
prev = &dn->dn_next;
}
write_unlock(&dn_lock);
}
int fcntl_dirnotify(int fd, struct file *filp, unsigned long arg) int fcntl_dirnotify(int fd, struct file *filp, unsigned long arg)
{ {
struct dnotify_struct *dn = NULL; struct dnotify_struct *dn;
struct dnotify_struct *odn; struct dnotify_struct *odn;
struct dnotify_struct **prev; struct dnotify_struct **prev;
struct inode *inode; struct inode *inode;
int turning_off = (arg & ~DN_MULTISHOT) == 0; fl_owner_t id = current->files;
if (!turning_off && !dir_notify_enable) if ((arg & ~DN_MULTISHOT) == 0) {
dnotify_flush(filp, id);
return 0;
}
if (!dir_notify_enable)
return -EINVAL; return -EINVAL;
inode = filp->f_dentry->d_inode; inode = filp->f_dentry->d_inode;
if (!S_ISDIR(inode->i_mode)) if (!S_ISDIR(inode->i_mode))
return -ENOTDIR; return -ENOTDIR;
if (!turning_off) { dn = kmem_cache_alloc(dn_cache, SLAB_KERNEL);
dn = kmem_cache_alloc(dn_cache, SLAB_KERNEL); if (dn == NULL)
if (dn == NULL) return -ENOMEM;
return -ENOMEM;
}
write_lock(&dn_lock); write_lock(&dn_lock);
prev = &inode->i_dnotify; prev = &inode->i_dnotify;
for (odn = *prev; odn != NULL; prev = &odn->dn_next, odn = *prev) while ((odn = *prev) != NULL) {
if ((odn->dn_owner == current->files) && (odn->dn_filp == filp)) if ((odn->dn_owner == id) && (odn->dn_filp == filp)) {
break; odn->dn_fd = fd;
if (odn != NULL) { odn->dn_mask |= arg;
if (turning_off) { inode->i_dnotify_mask |= arg & ~DN_MULTISHOT;
*prev = odn->dn_next; kmem_cache_free(dn_cache, dn);
redo_inode_mask(inode); goto out;
dn = odn;
goto out_free;
} }
odn->dn_fd = fd; prev = &odn->dn_next;
odn->dn_mask |= arg;
inode->i_dnotify_mask |= arg & ~DN_MULTISHOT;
goto out_free;
} }
if (turning_off)
goto out;
filp->f_owner.pid = current->pid; filp->f_owner.pid = current->pid;
filp->f_owner.uid = current->uid; filp->f_owner.uid = current->uid;
filp->f_owner.euid = current->euid; filp->f_owner.euid = current->euid;
dn->dn_magic = DNOTIFY_MAGIC;
dn->dn_mask = arg; dn->dn_mask = arg;
dn->dn_fd = fd; dn->dn_fd = fd;
dn->dn_filp = filp; dn->dn_filp = filp;
dn->dn_owner = current->files; dn->dn_owner = id;
inode->i_dnotify_mask |= arg & ~DN_MULTISHOT; inode->i_dnotify_mask |= arg & ~DN_MULTISHOT;
dn->dn_next = inode->i_dnotify; dn->dn_next = inode->i_dnotify;
inode->i_dnotify = dn; inode->i_dnotify = dn;
out: out:
write_unlock(&dn_lock); write_unlock(&dn_lock);
return 0; return 0;
out_free:
kmem_cache_free(dn_cache, dn);
goto out;
} }
void __inode_dir_notify(struct inode *inode, unsigned long event) void __inode_dir_notify(struct inode *inode, unsigned long event)
...@@ -104,11 +118,6 @@ void __inode_dir_notify(struct inode *inode, unsigned long event) ...@@ -104,11 +118,6 @@ void __inode_dir_notify(struct inode *inode, unsigned long event)
write_lock(&dn_lock); write_lock(&dn_lock);
prev = &inode->i_dnotify; prev = &inode->i_dnotify;
while ((dn = *prev) != NULL) { while ((dn = *prev) != NULL) {
if (dn->dn_magic != DNOTIFY_MAGIC) {
printk(KERN_ERR "__inode_dir_notify: bad magic "
"number in dnotify_struct!\n");
goto out;
}
if ((dn->dn_mask & event) == 0) { if ((dn->dn_mask & event) == 0) {
prev = &dn->dn_next; prev = &dn->dn_next;
continue; continue;
......
...@@ -1385,6 +1385,35 @@ int lmLogClose(struct super_block *sb, log_t * log) ...@@ -1385,6 +1385,35 @@ int lmLogClose(struct super_block *sb, log_t * log)
} }
/*
* NAME: lmLogWait()
*
* FUNCTION: wait for all outstanding log records to be written to disk
*/
void lmLogWait(log_t *log)
{
int i;
jFYI(1, ("lmLogWait: log:0x%p\n", log));
if (log->cqueue.head || !list_empty(&log->synclist)) {
/*
* If there was very recent activity, we may need to wait
* for the lazycommit thread to catch up
*/
for (i = 0; i < 800; i++) { /* Too much? */
current->state = TASK_INTERRUPTIBLE;
schedule_timeout(HZ / 4);
if ((log->cqueue.head == NULL) &&
list_empty(&log->synclist))
break;
}
}
assert(log->cqueue.head == NULL);
assert(list_empty(&log->synclist));
}
/* /*
* NAME: lmLogShutdown() * NAME: lmLogShutdown()
* *
...@@ -1411,23 +1440,7 @@ static int lmLogShutdown(log_t * log) ...@@ -1411,23 +1440,7 @@ static int lmLogShutdown(log_t * log)
jFYI(1, ("lmLogShutdown: log:0x%p\n", log)); jFYI(1, ("lmLogShutdown: log:0x%p\n", log));
if (log->cqueue.head || !list_empty(&log->synclist)) { lmLogWait(log);
/*
* If there was very recent activity, we may need to wait
* for the lazycommit thread to catch up
*/
int i;
for (i = 0; i < 800; i++) { /* Too much? */
current->state = TASK_INTERRUPTIBLE;
schedule_timeout(HZ / 4);
if ((log->cqueue.head == NULL) &&
list_empty(&log->synclist))
break;
}
}
assert(log->cqueue.head == NULL);
assert(list_empty(&log->synclist));
/* /*
* We need to make sure all of the "written" metapages * We need to make sure all of the "written" metapages
......
...@@ -489,6 +489,7 @@ typedef struct logsyncblk { ...@@ -489,6 +489,7 @@ typedef struct logsyncblk {
} }
extern int lmLogOpen(struct super_block *sb, log_t ** log); extern int lmLogOpen(struct super_block *sb, log_t ** log);
extern void lmLogWait(log_t * log);
extern int lmLogClose(struct super_block *sb, log_t * log); extern int lmLogClose(struct super_block *sb, log_t * log);
extern int lmLogSync(log_t * log, int nosyncwait); extern int lmLogSync(log_t * log, int nosyncwait);
extern int lmLogQuiesce(log_t * log); extern int lmLogQuiesce(log_t * log);
......
...@@ -64,15 +64,11 @@ int jfs_umount(struct super_block *sb) ...@@ -64,15 +64,11 @@ int jfs_umount(struct super_block *sb)
* *
* if mounted read-write and log based recovery was enabled * if mounted read-write and log based recovery was enabled
*/ */
if ((log = sbi->log)) { if ((log = sbi->log))
/* /*
* close log: * Wait for outstanding transactions to be written to log:
*
* remove file system from log active file system list.
*/ */
log = sbi->log; lmLogWait(log);
rc = lmLogClose(sb, log);
}
/* /*
* close fileset inode allocation map (aka fileset inode) * close fileset inode allocation map (aka fileset inode)
...@@ -112,6 +108,14 @@ int jfs_umount(struct super_block *sb) ...@@ -112,6 +108,14 @@ int jfs_umount(struct super_block *sb)
diFreeSpecial(ipbmap); diFreeSpecial(ipbmap);
sbi->ipimap = NULL; sbi->ipimap = NULL;
/*
* Make sure all metadata makes it to disk before we mark
* the superblock as clean
*/
filemap_fdatawait(sbi->direct_inode->i_mapping);
filemap_fdatawrite(sbi->direct_inode->i_mapping);
filemap_fdatawait(sbi->direct_inode->i_mapping);
/* /*
* ensure all file system file pages are propagated to their * ensure all file system file pages are propagated to their
* home blocks on disk (and their in-memory buffer pages are * home blocks on disk (and their in-memory buffer pages are
...@@ -120,10 +124,16 @@ int jfs_umount(struct super_block *sb) ...@@ -120,10 +124,16 @@ int jfs_umount(struct super_block *sb)
* consistent state) and log superblock active file system * consistent state) and log superblock active file system
* list (to signify skip logredo()). * list (to signify skip logredo()).
*/ */
if (log) /* log = NULL if read-only mount */ if (log) { /* log = NULL if read-only mount */
rc = updateSuper(sb, FM_CLEAN); rc = updateSuper(sb, FM_CLEAN);
/*
* close log:
*
* remove file system from log active file system list.
*/
rc = lmLogClose(sb, log);
}
jFYI(0, (" UnMount JFS Complete: %d\n", rc)); jFYI(0, (" UnMount JFS Complete: %d\n", rc));
return rc; return rc;
} }
...@@ -132,8 +142,9 @@ int jfs_umount(struct super_block *sb) ...@@ -132,8 +142,9 @@ int jfs_umount(struct super_block *sb)
int jfs_umount_rw(struct super_block *sb) int jfs_umount_rw(struct super_block *sb)
{ {
struct jfs_sb_info *sbi = JFS_SBI(sb); struct jfs_sb_info *sbi = JFS_SBI(sb);
log_t *log = sbi->log;
if (!sbi->log) if (!log)
return 0; return 0;
/* /*
...@@ -141,13 +152,19 @@ int jfs_umount_rw(struct super_block *sb) ...@@ -141,13 +152,19 @@ int jfs_umount_rw(struct super_block *sb)
* *
* remove file system from log active file system list. * remove file system from log active file system list.
*/ */
lmLogClose(sb, sbi->log); lmLogWait(log);
/*
* Make sure all metadata makes it to disk
*/
dbSync(sbi->ipbmap); dbSync(sbi->ipbmap);
diSync(sbi->ipimap); diSync(sbi->ipimap);
filemap_fdatawait(sbi->direct_inode->i_mapping);
filemap_fdatawrite(sbi->direct_inode->i_mapping);
filemap_fdatawait(sbi->direct_inode->i_mapping);
sbi->log = 0; updateSuper(sb, FM_CLEAN);
sbi->log = NULL;
return updateSuper(sb, FM_CLEAN); return lmLogClose(sb, log);
} }
...@@ -196,7 +196,7 @@ nfs_list_add_request(struct nfs_page *req, struct list_head *head) ...@@ -196,7 +196,7 @@ nfs_list_add_request(struct nfs_page *req, struct list_head *head)
BUG(); BUG();
} }
#endif #endif
for (pos = head->prev; pos != head; pos = pos->prev) { list_for_each_prev(pos, head) {
struct nfs_page *p = nfs_list_entry(pos); struct nfs_page *p = nfs_list_entry(pos);
if (page_index(p->wb_page) < pg_idx) if (page_index(p->wb_page) < pg_idx)
break; break;
......
...@@ -835,7 +835,7 @@ int filp_close(struct file *filp, fl_owner_t id) ...@@ -835,7 +835,7 @@ int filp_close(struct file *filp, fl_owner_t id)
retval = filp->f_op->flush(filp); retval = filp->f_op->flush(filp);
unlock_kernel(); unlock_kernel();
} }
fcntl_dirnotify(0, filp, 0); dnotify_flush(filp, id);
locks_remove_posix(filp, id); locks_remove_posix(filp, id);
fput(filp); fput(filp);
return retval; return retval;
......
...@@ -75,7 +75,8 @@ enum chipset_type { ...@@ -75,7 +75,8 @@ enum chipset_type {
ALI_GENERIC, ALI_GENERIC,
SVWRKS_HE, SVWRKS_HE,
SVWRKS_LE, SVWRKS_LE,
SVWRKS_GENERIC SVWRKS_GENERIC,
HP_ZX1,
}; };
typedef struct _agp_version { typedef struct _agp_version {
......
/* /*
* Directory notification for Linux * Directory notification for Linux
* *
* Copyright 2000 (C) Stephen Rothwell * Copyright (C) 2000,2002 Stephen Rothwell
*/ */
#include <linux/fs.h> #include <linux/fs.h>
struct dnotify_struct { struct dnotify_struct {
struct dnotify_struct * dn_next; struct dnotify_struct * dn_next;
int dn_magic;
unsigned long dn_mask; /* Events to be notified unsigned long dn_mask; /* Events to be notified
see linux/fcntl.h */ see linux/fcntl.h */
int dn_fd; int dn_fd;
...@@ -16,9 +15,8 @@ struct dnotify_struct { ...@@ -16,9 +15,8 @@ struct dnotify_struct {
fl_owner_t dn_owner; fl_owner_t dn_owner;
}; };
#define DNOTIFY_MAGIC 0x444E4F54
extern void __inode_dir_notify(struct inode *, unsigned long); extern void __inode_dir_notify(struct inode *, unsigned long);
extern void dnotify_flush(struct file *filp, fl_owner_t id);
extern int fcntl_dirnotify(int, struct file *, unsigned long); extern int fcntl_dirnotify(int, struct file *, unsigned long);
static inline void inode_dir_notify(struct inode *inode, unsigned long event) static inline void inode_dir_notify(struct inode *inode, unsigned long event)
......
...@@ -160,6 +160,7 @@ int access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, in ...@@ -160,6 +160,7 @@ int access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, in
put_page(page); put_page(page);
len -= bytes; len -= bytes;
buf += bytes; buf += bytes;
addr += bytes;
} }
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
mmput(mm); mmput(mm);
......
...@@ -1088,7 +1088,7 @@ asmlinkage long sys_swapon(const char * specialfile, int swap_flags) ...@@ -1088,7 +1088,7 @@ asmlinkage long sys_swapon(const char * specialfile, int swap_flags)
swap_list_unlock(); swap_list_unlock();
if (swap_map) if (swap_map)
vfree(swap_map); vfree(swap_map);
if (swap_file) if (swap_file && !IS_ERR(swap_file))
filp_close(swap_file, NULL); filp_close(swap_file, NULL);
out: out:
if (swap_header) if (swap_header)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment