2 * COPYRIGHT: See COPYING in the top level directory
3 * PROJECT: ReactOS kernel
4 * FILE: ntoskrnl/cc/view.c
5 * PURPOSE: Cache manager
7 * PROGRAMMERS: David Welch (welch@mcmail.com)
8 * Pierre Schweitzer (pierre@reactos.org)
11 /* NOTES **********************************************************************
13 * This is not the NT implementation of a file cache nor anything much like
16 * The general procedure for a filesystem to implement a read or write
17 * dispatch routine is as follows
19 * (1) If caching for the FCB hasn't been initiated then so do by calling
20 * CcInitializeFileCache.
22 * (2) For each 4k region which is being read or written obtain a cache page
23 * by calling CcRequestCachePage.
25 * (3) If either the page is being read or not completely written, and it is
26 * not up to date then read its data from the underlying medium. If the read
27 * fails then call CcReleaseCachePage with VALID as FALSE and return a error.
29 * (4) Copy the data into or out of the page as necessary.
31 * (5) Release the cache page
33 /* INCLUDES ******************************************************************/
39 #if defined (ALLOC_PRAGMA)
40 #pragma alloc_text(INIT, CcInitView)
43 /* GLOBALS *******************************************************************/
45 LIST_ENTRY DirtyVacbListHead
;
46 static LIST_ENTRY VacbLruListHead
;
48 NPAGED_LOOKASIDE_LIST iBcbLookasideList
;
49 static NPAGED_LOOKASIDE_LIST SharedCacheMapLookasideList
;
50 static NPAGED_LOOKASIDE_LIST VacbLookasideList
;
52 /* Internal vars (MS):
53 * - Threshold above which lazy writer will start action
54 * - Amount of dirty pages
55 * - List for deferred writes
56 * - Spinlock when dealing with the deferred list
57 * - List for "clean" shared cache maps
59 ULONG CcDirtyPageThreshold
= 0;
60 ULONG CcTotalDirtyPages
= 0;
61 LIST_ENTRY CcDeferredWrites
;
62 KSPIN_LOCK CcDeferredWriteSpinLock
;
63 LIST_ENTRY CcCleanSharedCacheMapList
;
66 ULONG
CcRosVacbIncRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
70 Refs
= InterlockedIncrement((PLONG
)&vacb
->ReferenceCount
);
71 if (vacb
->SharedCacheMap
->Trace
)
73 DbgPrint("(%s:%i) VACB %p ++RefCount=%lu, Dirty %u, PageOut %lu\n",
74 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
79 ULONG
CcRosVacbDecRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
83 Refs
= InterlockedDecrement((PLONG
)&vacb
->ReferenceCount
);
84 ASSERT(!(Refs
== 0 && vacb
->Dirty
));
85 if (vacb
->SharedCacheMap
->Trace
)
87 DbgPrint("(%s:%i) VACB %p --RefCount=%lu, Dirty %u, PageOut %lu\n",
88 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
93 CcRosInternalFreeVacb(vacb
);
98 ULONG
CcRosVacbGetRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
102 Refs
= InterlockedCompareExchange((PLONG
)&vacb
->ReferenceCount
, 0, 0);
103 if (vacb
->SharedCacheMap
->Trace
)
105 DbgPrint("(%s:%i) VACB %p ==RefCount=%lu, Dirty %u, PageOut %lu\n",
106 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
114 /* FUNCTIONS *****************************************************************/
119 PROS_SHARED_CACHE_MAP SharedCacheMap
,
124 PLIST_ENTRY current_entry
;
130 SharedCacheMap
->Trace
= Trace
;
134 DPRINT1("Enabling Tracing for CacheMap 0x%p:\n", SharedCacheMap
);
136 oldirql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
137 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
139 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
140 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
142 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
143 current_entry
= current_entry
->Flink
;
145 DPRINT1(" VACB 0x%p enabled, RefCount %lu, Dirty %u, PageOut %lu\n",
146 current
, current
->ReferenceCount
, current
->Dirty
, current
->PageOut
);
149 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
150 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldirql
);
154 DPRINT1("Disabling Tracing for CacheMap 0x%p:\n", SharedCacheMap
);
158 UNREFERENCED_PARAMETER(SharedCacheMap
);
159 UNREFERENCED_PARAMETER(Trace
);
170 CcRosUnmarkDirtyVacb(Vacb
, TRUE
);
172 Status
= CcWriteVirtualAddress(Vacb
);
173 if (!NT_SUCCESS(Status
))
175 CcRosMarkDirtyVacb(Vacb
);
183 CcRosFlushDirtyPages (
187 BOOLEAN CalledFromLazy
)
189 PLIST_ENTRY current_entry
;
195 DPRINT("CcRosFlushDirtyPages(Target %lu)\n", Target
);
199 KeEnterCriticalRegion();
200 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
202 current_entry
= DirtyVacbListHead
.Flink
;
203 if (current_entry
== &DirtyVacbListHead
)
205 DPRINT("No Dirty pages\n");
208 while ((current_entry
!= &DirtyVacbListHead
) && (Target
> 0))
210 current
= CONTAINING_RECORD(current_entry
,
213 current_entry
= current_entry
->Flink
;
215 CcRosVacbIncRefCount(current
);
217 /* When performing lazy write, don't handle temporary files */
218 if (CalledFromLazy
&&
219 BooleanFlagOn(current
->SharedCacheMap
->FileObject
->Flags
, FO_TEMPORARY_FILE
))
221 CcRosVacbDecRefCount(current
);
225 ASSERT(current
->Dirty
);
227 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
229 Locked
= current
->SharedCacheMap
->Callbacks
->AcquireForLazyWrite(
230 current
->SharedCacheMap
->LazyWriteContext
, Wait
);
233 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
234 CcRosVacbDecRefCount(current
);
238 Status
= CcRosFlushVacb(current
);
240 current
->SharedCacheMap
->Callbacks
->ReleaseFromLazyWrite(
241 current
->SharedCacheMap
->LazyWriteContext
);
243 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
244 CcRosVacbDecRefCount(current
);
246 if (!NT_SUCCESS(Status
) && (Status
!= STATUS_END_OF_FILE
) &&
247 (Status
!= STATUS_MEDIA_WRITE_PROTECTED
))
249 DPRINT1("CC: Failed to flush VACB.\n");
255 /* How many pages did we free? */
256 PagesFreed
= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
257 (*Count
) += PagesFreed
;
259 /* Make sure we don't overflow target! */
260 if (Target
< PagesFreed
)
262 /* If we would have, jump to zero directly */
267 Target
-= PagesFreed
;
271 current_entry
= DirtyVacbListHead
.Flink
;
274 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
275 KeLeaveCriticalRegion();
277 DPRINT("CcRosFlushDirtyPages() finished\n");
278 return STATUS_SUCCESS
;
287 * FUNCTION: Try to free some memory from the file cache.
289 * Target - The number of pages to be freed.
290 * Priority - The priority of free (currently unused).
291 * NrFreed - Points to a variable where the number of pages
292 * actually freed is returned.
295 PLIST_ENTRY current_entry
;
302 BOOLEAN FlushedPages
= FALSE
;
304 DPRINT("CcRosTrimCache(Target %lu)\n", Target
);
306 InitializeListHead(&FreeList
);
311 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
313 current_entry
= VacbLruListHead
.Flink
;
314 while (current_entry
!= &VacbLruListHead
)
318 current
= CONTAINING_RECORD(current_entry
,
321 current_entry
= current_entry
->Flink
;
323 KeAcquireSpinLockAtDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
325 /* Reference the VACB */
326 CcRosVacbIncRefCount(current
);
328 /* Check if it's mapped and not dirty */
329 if (InterlockedCompareExchange((PLONG
)¤t
->MappedCount
, 0, 0) > 0 && !current
->Dirty
)
331 /* We have to break these locks because Cc sucks */
332 KeReleaseSpinLockFromDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
333 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
335 /* Page out the VACB */
336 for (i
= 0; i
< VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
; i
++)
338 Page
= (PFN_NUMBER
)(MmGetPhysicalAddress((PUCHAR
)current
->BaseAddress
+ (i
* PAGE_SIZE
)).QuadPart
>> PAGE_SHIFT
);
340 MmPageOutPhysicalAddress(Page
);
343 /* Reacquire the locks */
344 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
345 KeAcquireSpinLockAtDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
348 /* Dereference the VACB */
349 Refs
= CcRosVacbDecRefCount(current
);
351 /* Check if we can free this entry now */
354 ASSERT(!current
->Dirty
);
355 ASSERT(!current
->MappedCount
);
358 RemoveEntryList(¤t
->CacheMapVacbListEntry
);
359 RemoveEntryList(¤t
->VacbLruListEntry
);
360 InitializeListHead(¤t
->VacbLruListEntry
);
361 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
363 /* Calculate how many pages we freed for Mm */
364 PagesFreed
= min(VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
, Target
);
365 Target
-= PagesFreed
;
366 (*NrFreed
) += PagesFreed
;
369 KeReleaseSpinLockFromDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
372 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
374 /* Try flushing pages if we haven't met our target */
375 if ((Target
> 0) && !FlushedPages
)
377 /* Flush dirty pages to disk */
378 CcRosFlushDirtyPages(Target
, &PagesFreed
, FALSE
, FALSE
);
381 /* We can only swap as many pages as we flushed */
382 if (PagesFreed
< Target
) Target
= PagesFreed
;
384 /* Check if we flushed anything */
387 /* Try again after flushing dirty pages */
388 DPRINT("Flushed %lu dirty cache pages to disk\n", PagesFreed
);
393 while (!IsListEmpty(&FreeList
))
397 current_entry
= RemoveHeadList(&FreeList
);
398 current
= CONTAINING_RECORD(current_entry
,
400 CacheMapVacbListEntry
);
401 InitializeListHead(¤t
->CacheMapVacbListEntry
);
402 Refs
= CcRosVacbDecRefCount(current
);
406 DPRINT("Evicted %lu cache pages\n", (*NrFreed
));
408 return STATUS_SUCCESS
;
414 PROS_SHARED_CACHE_MAP SharedCacheMap
,
421 ASSERT(SharedCacheMap
);
423 DPRINT("CcRosReleaseVacb(SharedCacheMap 0x%p, Vacb 0x%p, Valid %u)\n",
424 SharedCacheMap
, Vacb
, Valid
);
428 if (Dirty
&& !Vacb
->Dirty
)
430 CcRosMarkDirtyVacb(Vacb
);
435 if (InterlockedIncrement((PLONG
)&Vacb
->MappedCount
) == 1)
437 CcRosVacbIncRefCount(Vacb
);
441 Refs
= CcRosVacbDecRefCount(Vacb
);
444 return STATUS_SUCCESS
;
447 /* Returns with VACB Lock Held! */
451 PROS_SHARED_CACHE_MAP SharedCacheMap
,
454 PLIST_ENTRY current_entry
;
458 ASSERT(SharedCacheMap
);
460 DPRINT("CcRosLookupVacb(SharedCacheMap 0x%p, FileOffset %I64u)\n",
461 SharedCacheMap
, FileOffset
);
463 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
464 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
466 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
467 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
469 current
= CONTAINING_RECORD(current_entry
,
471 CacheMapVacbListEntry
);
472 if (IsPointInRange(current
->FileOffset
.QuadPart
,
473 VACB_MAPPING_GRANULARITY
,
476 CcRosVacbIncRefCount(current
);
477 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
478 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
481 if (current
->FileOffset
.QuadPart
> FileOffset
)
483 current_entry
= current_entry
->Flink
;
486 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
487 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
498 PROS_SHARED_CACHE_MAP SharedCacheMap
;
500 SharedCacheMap
= Vacb
->SharedCacheMap
;
502 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
503 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
505 ASSERT(!Vacb
->Dirty
);
507 InsertTailList(&DirtyVacbListHead
, &Vacb
->DirtyVacbListEntry
);
508 CcTotalDirtyPages
+= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
509 Vacb
->SharedCacheMap
->DirtyPages
+= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
510 CcRosVacbIncRefCount(Vacb
);
512 /* Move to the tail of the LRU list */
513 RemoveEntryList(&Vacb
->VacbLruListEntry
);
514 InsertTailList(&VacbLruListHead
, &Vacb
->VacbLruListEntry
);
518 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
520 /* Schedule a lazy writer run to now that we have dirty VACB */
521 if (!LazyWriter
.ScanActive
)
523 CcScheduleLazyWriteScan(FALSE
);
525 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
530 CcRosUnmarkDirtyVacb (
535 PROS_SHARED_CACHE_MAP SharedCacheMap
;
537 SharedCacheMap
= Vacb
->SharedCacheMap
;
541 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
542 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
549 RemoveEntryList(&Vacb
->DirtyVacbListEntry
);
550 InitializeListHead(&Vacb
->DirtyVacbListEntry
);
551 CcTotalDirtyPages
-= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
552 Vacb
->SharedCacheMap
->DirtyPages
-= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
553 CcRosVacbDecRefCount(Vacb
);
557 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
558 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
565 PROS_SHARED_CACHE_MAP SharedCacheMap
,
570 ASSERT(SharedCacheMap
);
572 DPRINT("CcRosMarkDirtyVacb(SharedCacheMap 0x%p, FileOffset %I64u)\n",
573 SharedCacheMap
, FileOffset
);
575 Vacb
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
578 KeBugCheck(CACHE_MANAGER
);
581 CcRosReleaseVacb(SharedCacheMap
, Vacb
, Vacb
->Valid
, TRUE
, FALSE
);
583 return STATUS_SUCCESS
;
587 * Note: this is not the contrary function of
588 * CcRosMapVacbInKernelSpace()
593 PROS_SHARED_CACHE_MAP SharedCacheMap
,
599 ASSERT(SharedCacheMap
);
601 DPRINT("CcRosUnmapVacb(SharedCacheMap 0x%p, FileOffset %I64u, NowDirty %u)\n",
602 SharedCacheMap
, FileOffset
, NowDirty
);
604 Vacb
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
607 return STATUS_UNSUCCESSFUL
;
610 ASSERT(Vacb
->MappedCount
!= 0);
611 if (InterlockedDecrement((PLONG
)&Vacb
->MappedCount
) == 0)
613 CcRosVacbDecRefCount(Vacb
);
616 CcRosReleaseVacb(SharedCacheMap
, Vacb
, Vacb
->Valid
, NowDirty
, FALSE
);
618 return STATUS_SUCCESS
;
623 CcRosMapVacbInKernelSpace(
628 ULONG_PTR NumberOfPages
;
629 PVOID BaseAddress
= NULL
;
631 /* Create a memory area. */
632 MmLockAddressSpace(MmGetKernelAddressSpace());
633 Status
= MmCreateMemoryArea(MmGetKernelAddressSpace(),
634 0, // nothing checks for VACB mareas, so set to 0
636 VACB_MAPPING_GRANULARITY
,
638 (PMEMORY_AREA
*)&Vacb
->MemoryArea
,
641 ASSERT(Vacb
->BaseAddress
== NULL
);
642 Vacb
->BaseAddress
= BaseAddress
;
643 MmUnlockAddressSpace(MmGetKernelAddressSpace());
644 if (!NT_SUCCESS(Status
))
646 DPRINT1("MmCreateMemoryArea failed with %lx for VACB %p\n", Status
, Vacb
);
650 ASSERT(((ULONG_PTR
)Vacb
->BaseAddress
% PAGE_SIZE
) == 0);
651 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
> (ULONG_PTR
)MmSystemRangeStart
);
652 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ VACB_MAPPING_GRANULARITY
- 1 > (ULONG_PTR
)MmSystemRangeStart
);
654 /* Create a virtual mapping for this memory area */
655 NumberOfPages
= BYTES_TO_PAGES(VACB_MAPPING_GRANULARITY
);
656 for (i
= 0; i
< NumberOfPages
; i
++)
658 PFN_NUMBER PageFrameNumber
;
660 MI_SET_USAGE(MI_USAGE_CACHE
);
661 Status
= MmRequestPageMemoryConsumer(MC_CACHE
, TRUE
, &PageFrameNumber
);
662 if (PageFrameNumber
== 0)
664 DPRINT1("Unable to allocate page\n");
665 KeBugCheck(MEMORY_MANAGEMENT
);
668 ASSERT(BaseAddress
== Vacb
->BaseAddress
);
669 ASSERT(i
* PAGE_SIZE
< VACB_MAPPING_GRANULARITY
);
670 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
) >= (ULONG_PTR
)BaseAddress
);
671 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
) > (ULONG_PTR
)MmSystemRangeStart
);
673 Status
= MmCreateVirtualMapping(NULL
,
674 (PVOID
)((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
)),
678 if (!NT_SUCCESS(Status
))
680 DPRINT1("Unable to create virtual mapping\n");
681 KeBugCheck(MEMORY_MANAGEMENT
);
685 return STATUS_SUCCESS
;
690 CcRosFreeUnusedVacb (
698 PLIST_ENTRY current_entry
;
702 InitializeListHead(&FreeList
);
704 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
706 /* Browse all the available VACB */
707 current_entry
= VacbLruListHead
.Flink
;
708 while (current_entry
!= &VacbLruListHead
)
712 current
= CONTAINING_RECORD(current_entry
,
715 current_entry
= current_entry
->Flink
;
717 KeAcquireSpinLockAtDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
719 /* Only deal with unused VACB, we will free them */
720 Refs
= CcRosVacbGetRefCount(current
);
723 ASSERT(!current
->Dirty
);
724 ASSERT(!current
->MappedCount
);
727 /* Reset and move to free list */
728 RemoveEntryList(¤t
->CacheMapVacbListEntry
);
729 RemoveEntryList(¤t
->VacbLruListEntry
);
730 InitializeListHead(¤t
->VacbLruListEntry
);
731 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
734 KeReleaseSpinLockFromDpcLevel(¤t
->SharedCacheMap
->CacheMapLock
);
738 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
740 /* And now, free any of the found VACB, that'll free memory! */
741 while (!IsListEmpty(&FreeList
))
745 current_entry
= RemoveHeadList(&FreeList
);
746 current
= CONTAINING_RECORD(current_entry
,
748 CacheMapVacbListEntry
);
749 InitializeListHead(¤t
->CacheMapVacbListEntry
);
750 Refs
= CcRosVacbDecRefCount(current
);
755 /* If we freed at least one VACB, return success */
761 /* If caller asked for free count, return it */
773 PROS_SHARED_CACHE_MAP SharedCacheMap
,
779 PLIST_ENTRY current_entry
;
785 ASSERT(SharedCacheMap
);
787 DPRINT("CcRosCreateVacb()\n");
789 if (FileOffset
>= SharedCacheMap
->SectionSize
.QuadPart
)
792 return STATUS_INVALID_PARAMETER
;
795 current
= ExAllocateFromNPagedLookasideList(&VacbLookasideList
);
796 current
->BaseAddress
= NULL
;
797 current
->Valid
= FALSE
;
798 current
->Dirty
= FALSE
;
799 current
->PageOut
= FALSE
;
800 current
->FileOffset
.QuadPart
= ROUND_DOWN(FileOffset
, VACB_MAPPING_GRANULARITY
);
801 current
->SharedCacheMap
= SharedCacheMap
;
803 if (SharedCacheMap
->Trace
)
805 DPRINT1("CacheMap 0x%p: new VACB: 0x%p\n", SharedCacheMap
, current
);
808 current
->MappedCount
= 0;
809 current
->ReferenceCount
= 0;
810 InitializeListHead(¤t
->CacheMapVacbListEntry
);
811 InitializeListHead(¤t
->DirtyVacbListEntry
);
812 InitializeListHead(¤t
->VacbLruListEntry
);
814 CcRosVacbIncRefCount(current
);
818 /* Map VACB in kernel space */
819 Status
= CcRosMapVacbInKernelSpace(current
);
820 if (!NT_SUCCESS(Status
))
823 /* If no space left, try to prune unused VACB
824 * to recover space to map our VACB
825 * If it succeed, retry to map, otherwise
828 if (!Retried
&& CcRosFreeUnusedVacb(&Freed
))
830 DPRINT("Prunned %d VACB, trying again\n", Freed
);
835 ExFreeToNPagedLookasideList(&VacbLookasideList
, current
);
839 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
842 /* There is window between the call to CcRosLookupVacb
843 * and CcRosCreateVacb. We must check if a VACB for the
844 * file offset exist. If there is a VACB, we release
845 * our newly created VACB and return the existing one.
847 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
848 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
850 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
852 current
= CONTAINING_RECORD(current_entry
,
854 CacheMapVacbListEntry
);
855 if (IsPointInRange(current
->FileOffset
.QuadPart
,
856 VACB_MAPPING_GRANULARITY
,
859 CcRosVacbIncRefCount(current
);
860 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
862 if (SharedCacheMap
->Trace
)
864 DPRINT1("CacheMap 0x%p: deleting newly created VACB 0x%p ( found existing one 0x%p )\n",
870 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
872 Refs
= CcRosVacbDecRefCount(*Vacb
);
876 return STATUS_SUCCESS
;
878 if (current
->FileOffset
.QuadPart
< FileOffset
)
880 ASSERT(previous
== NULL
||
881 previous
->FileOffset
.QuadPart
< current
->FileOffset
.QuadPart
);
884 if (current
->FileOffset
.QuadPart
> FileOffset
)
886 current_entry
= current_entry
->Flink
;
888 /* There was no existing VACB. */
892 InsertHeadList(&previous
->CacheMapVacbListEntry
, ¤t
->CacheMapVacbListEntry
);
896 InsertHeadList(&SharedCacheMap
->CacheMapVacbListHead
, ¤t
->CacheMapVacbListEntry
);
898 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
899 InsertTailList(&VacbLruListHead
, ¤t
->VacbLruListEntry
);
900 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
902 MI_SET_USAGE(MI_USAGE_CACHE
);
904 if ((SharedCacheMap
->FileObject
) && (SharedCacheMap
->FileObject
->FileName
.Buffer
))
908 pos
= wcsrchr(SharedCacheMap
->FileObject
->FileName
.Buffer
, '\\');
911 len
= wcslen(pos
) * sizeof(WCHAR
);
912 snprintf(MI_PFN_CURRENT_PROCESS_NAME
, min(16, len
), "%S", pos
);
916 snprintf(MI_PFN_CURRENT_PROCESS_NAME
, min(16, len
), "%wZ", &SharedCacheMap
->FileObject
->FileName
);
921 /* Reference it to allow release */
922 CcRosVacbIncRefCount(current
);
930 PROS_SHARED_CACHE_MAP SharedCacheMap
,
932 PLONGLONG BaseOffset
,
942 ASSERT(SharedCacheMap
);
944 DPRINT("CcRosGetVacb()\n");
947 * Look for a VACB already mapping the same data.
949 current
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
953 * Otherwise create a new VACB.
955 Status
= CcRosCreateVacb(SharedCacheMap
, FileOffset
, ¤t
);
956 if (!NT_SUCCESS(Status
))
962 Refs
= CcRosVacbGetRefCount(current
);
964 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
966 /* Move to the tail of the LRU list */
967 RemoveEntryList(¤t
->VacbLruListEntry
);
968 InsertTailList(&VacbLruListHead
, ¤t
->VacbLruListEntry
);
970 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
973 * Return information about the VACB to the caller.
975 *UptoDate
= current
->Valid
;
976 *BaseAddress
= current
->BaseAddress
;
977 DPRINT("*BaseAddress %p\n", *BaseAddress
);
979 *BaseOffset
= current
->FileOffset
.QuadPart
;
983 return STATUS_SUCCESS
;
989 PROS_SHARED_CACHE_MAP SharedCacheMap
,
995 * FUNCTION: Request a page mapping for a shared cache map
1000 ASSERT(SharedCacheMap
);
1002 if (FileOffset
% VACB_MAPPING_GRANULARITY
!= 0)
1004 DPRINT1("Bad fileoffset %I64x should be multiple of %x",
1005 FileOffset
, VACB_MAPPING_GRANULARITY
);
1006 KeBugCheck(CACHE_MANAGER
);
1009 return CcRosGetVacb(SharedCacheMap
,
1021 MEMORY_AREA
* MemoryArea
,
1024 SWAPENTRY SwapEntry
,
1027 ASSERT(SwapEntry
== 0);
1030 ASSERT(MmGetReferenceCountPage(Page
) == 1);
1031 MmReleasePageMemoryConsumer(MC_CACHE
, Page
);
1036 CcRosInternalFreeVacb (
1039 * FUNCTION: Releases a VACB associated with a shared cache map
1042 DPRINT("Freeing VACB 0x%p\n", Vacb
);
1044 if (Vacb
->SharedCacheMap
->Trace
)
1046 DPRINT1("CacheMap 0x%p: deleting VACB: 0x%p\n", Vacb
->SharedCacheMap
, Vacb
);
1050 MmLockAddressSpace(MmGetKernelAddressSpace());
1051 MmFreeMemoryArea(MmGetKernelAddressSpace(),
1055 MmUnlockAddressSpace(MmGetKernelAddressSpace());
1057 if (Vacb
->ReferenceCount
!= 0)
1059 DPRINT1("Invalid free: %ld\n", Vacb
->ReferenceCount
);
1060 if (Vacb
->SharedCacheMap
->FileObject
&& Vacb
->SharedCacheMap
->FileObject
->FileName
.Length
)
1062 DPRINT1("For file: %wZ\n", &Vacb
->SharedCacheMap
->FileObject
->FileName
);
1066 ASSERT(Vacb
->ReferenceCount
== 0);
1067 ASSERT(IsListEmpty(&Vacb
->CacheMapVacbListEntry
));
1068 ASSERT(IsListEmpty(&Vacb
->DirtyVacbListEntry
));
1069 ASSERT(IsListEmpty(&Vacb
->VacbLruListEntry
));
1070 RtlFillMemory(Vacb
, sizeof(*Vacb
), 0xfd);
1071 ExFreeToNPagedLookasideList(&VacbLookasideList
, Vacb
);
1072 return STATUS_SUCCESS
;
1081 IN PSECTION_OBJECT_POINTERS SectionObjectPointers
,
1082 IN PLARGE_INTEGER FileOffset OPTIONAL
,
1084 OUT PIO_STATUS_BLOCK IoStatus
)
1086 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1087 LARGE_INTEGER Offset
;
1088 LONGLONG RemainingLength
;
1092 CCTRACE(CC_API_DEBUG
, "SectionObjectPointers=%p FileOffset=%p Length=%lu\n",
1093 SectionObjectPointers
, FileOffset
, Length
);
1095 DPRINT("CcFlushCache(SectionObjectPointers 0x%p, FileOffset 0x%p, Length %lu, IoStatus 0x%p)\n",
1096 SectionObjectPointers
, FileOffset
, Length
, IoStatus
);
1098 if (SectionObjectPointers
&& SectionObjectPointers
->SharedCacheMap
)
1100 SharedCacheMap
= SectionObjectPointers
->SharedCacheMap
;
1101 ASSERT(SharedCacheMap
);
1104 Offset
= *FileOffset
;
1105 RemainingLength
= Length
;
1109 Offset
.QuadPart
= 0;
1110 RemainingLength
= SharedCacheMap
->FileSize
.QuadPart
;
1115 IoStatus
->Status
= STATUS_SUCCESS
;
1116 IoStatus
->Information
= 0;
1119 while (RemainingLength
> 0)
1121 current
= CcRosLookupVacb(SharedCacheMap
, Offset
.QuadPart
);
1122 if (current
!= NULL
)
1126 Status
= CcRosFlushVacb(current
);
1127 if (!NT_SUCCESS(Status
) && IoStatus
!= NULL
)
1129 IoStatus
->Status
= Status
;
1133 CcRosReleaseVacb(SharedCacheMap
, current
, current
->Valid
, current
->Dirty
, FALSE
);
1136 Offset
.QuadPart
+= VACB_MAPPING_GRANULARITY
;
1137 RemainingLength
-= min(RemainingLength
, VACB_MAPPING_GRANULARITY
);
1144 IoStatus
->Status
= STATUS_INVALID_PARAMETER
;
1151 CcRosDeleteFileCache (
1152 PFILE_OBJECT FileObject
,
1153 PROS_SHARED_CACHE_MAP SharedCacheMap
,
1156 * FUNCTION: Releases the shared cache map associated with a file object
1159 PLIST_ENTRY current_entry
;
1161 LIST_ENTRY FreeList
;
1163 ASSERT(SharedCacheMap
);
1165 SharedCacheMap
->OpenCount
++;
1166 KeReleaseQueuedSpinLock(LockQueueMasterLock
, *OldIrql
);
1168 CcFlushCache(FileObject
->SectionObjectPointer
, NULL
, 0, NULL
);
1170 *OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1171 SharedCacheMap
->OpenCount
--;
1172 if (SharedCacheMap
->OpenCount
== 0)
1174 FileObject
->SectionObjectPointer
->SharedCacheMap
= NULL
;
1179 InitializeListHead(&FreeList
);
1180 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
1181 while (!IsListEmpty(&SharedCacheMap
->CacheMapVacbListHead
))
1183 current_entry
= RemoveTailList(&SharedCacheMap
->CacheMapVacbListHead
);
1184 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
1186 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
1187 RemoveEntryList(¤t
->VacbLruListEntry
);
1188 InitializeListHead(¤t
->VacbLruListEntry
);
1191 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
1192 CcRosUnmarkDirtyVacb(current
, FALSE
);
1193 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
1194 DPRINT1("Freeing dirty VACB\n");
1196 if (current
->MappedCount
!= 0)
1198 current
->MappedCount
= 0;
1199 NT_VERIFY(CcRosVacbDecRefCount(current
) > 0);
1200 DPRINT1("Freeing mapped VACB\n");
1202 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
1204 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
1207 SharedCacheMap
->Trace
= FALSE
;
1209 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
1211 KeReleaseQueuedSpinLock(LockQueueMasterLock
, *OldIrql
);
1212 ObDereferenceObject(SharedCacheMap
->FileObject
);
1214 while (!IsListEmpty(&FreeList
))
1218 current_entry
= RemoveTailList(&FreeList
);
1219 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
1220 InitializeListHead(¤t
->CacheMapVacbListEntry
);
1221 Refs
= CcRosVacbDecRefCount(current
);
1222 #if DBG // CORE-14578
1225 DPRINT1("Leaking VACB %p attached to %p (%I64d)\n", current
, FileObject
, current
->FileOffset
.QuadPart
);
1226 DPRINT1("There are: %d references left\n", Refs
);
1227 DPRINT1("Map: %d\n", current
->MappedCount
);
1228 DPRINT1("Dirty: %d\n", current
->Dirty
);
1229 if (FileObject
->FileName
.Length
!= 0)
1231 DPRINT1("File was: %wZ\n", &FileObject
->FileName
);
1233 else if (FileObject
->FsContext
!= NULL
&&
1234 ((PFSRTL_COMMON_FCB_HEADER
)(FileObject
->FsContext
))->NodeTypeCode
== 0x0502 &&
1235 ((PFSRTL_COMMON_FCB_HEADER
)(FileObject
->FsContext
))->NodeByteSize
== 0x1F8 &&
1236 ((PUNICODE_STRING
)(((PUCHAR
)FileObject
->FsContext
) + 0x100))->Length
!= 0)
1238 DPRINT1("File was: %wZ (FastFAT)\n", (PUNICODE_STRING
)(((PUCHAR
)FileObject
->FsContext
) + 0x100));
1242 DPRINT1("No name for the file\n");
1250 *OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1251 RemoveEntryList(&SharedCacheMap
->SharedCacheMapLinks
);
1252 KeReleaseQueuedSpinLock(LockQueueMasterLock
, *OldIrql
);
1254 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1255 *OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1257 return STATUS_SUCCESS
;
1262 CcRosReferenceCache (
1263 PFILE_OBJECT FileObject
)
1265 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1268 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1269 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1270 ASSERT(SharedCacheMap
);
1271 ASSERT(SharedCacheMap
->OpenCount
!= 0);
1272 SharedCacheMap
->OpenCount
++;
1273 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1278 CcRosRemoveIfClosed (
1279 PSECTION_OBJECT_POINTERS SectionObjectPointer
)
1281 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1284 DPRINT("CcRosRemoveIfClosed()\n");
1285 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1286 SharedCacheMap
= SectionObjectPointer
->SharedCacheMap
;
1287 if (SharedCacheMap
&& SharedCacheMap
->OpenCount
== 0)
1289 CcRosDeleteFileCache(SharedCacheMap
->FileObject
, SharedCacheMap
, &OldIrql
);
1291 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1297 CcRosDereferenceCache (
1298 PFILE_OBJECT FileObject
)
1300 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1303 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1304 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1305 ASSERT(SharedCacheMap
);
1306 if (SharedCacheMap
->OpenCount
> 0)
1308 SharedCacheMap
->OpenCount
--;
1309 if (SharedCacheMap
->OpenCount
== 0)
1311 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1312 MmFreeSectionSegments(SharedCacheMap
->FileObject
);
1314 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1315 CcRosDeleteFileCache(FileObject
, SharedCacheMap
, &OldIrql
);
1316 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1321 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1326 CcRosReleaseFileCache (
1327 PFILE_OBJECT FileObject
)
1329 * FUNCTION: Called by the file system when a handle to a file object
1334 PPRIVATE_CACHE_MAP PrivateMap
;
1335 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1337 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1339 if (FileObject
->SectionObjectPointer
->SharedCacheMap
!= NULL
)
1341 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1343 /* Closing the handle, so kill the private cache map
1344 * Before you event try to remove it from FO, always
1345 * lock the master lock, to be sure not to race
1346 * with a potential read ahead ongoing!
1348 PrivateMap
= FileObject
->PrivateCacheMap
;
1349 FileObject
->PrivateCacheMap
= NULL
;
1351 if (PrivateMap
!= NULL
)
1353 /* Remove it from the file */
1354 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
1355 RemoveEntryList(&PrivateMap
->PrivateLinks
);
1356 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
1359 if (PrivateMap
!= &SharedCacheMap
->PrivateCacheMap
)
1361 ExFreePoolWithTag(PrivateMap
, TAG_PRIVATE_CACHE_MAP
);
1365 PrivateMap
->NodeTypeCode
= 0;
1368 if (SharedCacheMap
->OpenCount
> 0)
1370 SharedCacheMap
->OpenCount
--;
1371 if (SharedCacheMap
->OpenCount
== 0)
1373 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1374 MmFreeSectionSegments(SharedCacheMap
->FileObject
);
1376 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1377 CcRosDeleteFileCache(FileObject
, SharedCacheMap
, &OldIrql
);
1378 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1380 return STATUS_SUCCESS
;
1385 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1386 return STATUS_SUCCESS
;
1391 CcRosInitializeFileCache (
1392 PFILE_OBJECT FileObject
,
1393 PCC_FILE_SIZES FileSizes
,
1395 PCACHE_MANAGER_CALLBACKS CallBacks
,
1396 PVOID LazyWriterContext
)
1398 * FUNCTION: Initializes a shared cache map for a file object
1403 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1405 DPRINT("CcRosInitializeFileCache(FileObject 0x%p)\n", FileObject
);
1408 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1409 if (SharedCacheMap
== NULL
)
1412 SharedCacheMap
= ExAllocateFromNPagedLookasideList(&SharedCacheMapLookasideList
);
1413 if (SharedCacheMap
== NULL
)
1415 return STATUS_INSUFFICIENT_RESOURCES
;
1417 RtlZeroMemory(SharedCacheMap
, sizeof(*SharedCacheMap
));
1418 SharedCacheMap
->NodeTypeCode
= NODE_TYPE_SHARED_MAP
;
1419 SharedCacheMap
->NodeByteSize
= sizeof(*SharedCacheMap
);
1420 SharedCacheMap
->FileObject
= FileObject
;
1421 SharedCacheMap
->Callbacks
= CallBacks
;
1422 SharedCacheMap
->LazyWriteContext
= LazyWriterContext
;
1423 SharedCacheMap
->SectionSize
= FileSizes
->AllocationSize
;
1424 SharedCacheMap
->FileSize
= FileSizes
->FileSize
;
1425 SharedCacheMap
->PinAccess
= PinAccess
;
1426 SharedCacheMap
->DirtyPageThreshold
= 0;
1427 SharedCacheMap
->DirtyPages
= 0;
1428 InitializeListHead(&SharedCacheMap
->PrivateList
);
1429 KeInitializeSpinLock(&SharedCacheMap
->CacheMapLock
);
1430 InitializeListHead(&SharedCacheMap
->CacheMapVacbListHead
);
1431 InitializeListHead(&SharedCacheMap
->BcbList
);
1434 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1437 if (FileObject
->SectionObjectPointer
->SharedCacheMap
== NULL
)
1439 ObReferenceObjectByPointer(FileObject
,
1443 FileObject
->SectionObjectPointer
->SharedCacheMap
= SharedCacheMap
;
1445 InsertTailList(&CcCleanSharedCacheMapList
, &SharedCacheMap
->SharedCacheMapLinks
);
1449 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1450 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1453 if (FileObject
->PrivateCacheMap
== NULL
)
1455 PPRIVATE_CACHE_MAP PrivateMap
;
1457 /* Allocate the private cache map for this handle */
1458 if (SharedCacheMap
->PrivateCacheMap
.NodeTypeCode
!= 0)
1460 PrivateMap
= ExAllocatePoolWithTag(NonPagedPool
, sizeof(PRIVATE_CACHE_MAP
), TAG_PRIVATE_CACHE_MAP
);
1464 PrivateMap
= &SharedCacheMap
->PrivateCacheMap
;
1467 if (PrivateMap
== NULL
)
1469 /* If we also allocated the shared cache map for this file, kill it */
1472 RemoveEntryList(&SharedCacheMap
->SharedCacheMapLinks
);
1474 FileObject
->SectionObjectPointer
->SharedCacheMap
= NULL
;
1475 ObDereferenceObject(FileObject
);
1476 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1479 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1480 return STATUS_INSUFFICIENT_RESOURCES
;
1484 RtlZeroMemory(PrivateMap
, sizeof(PRIVATE_CACHE_MAP
));
1485 PrivateMap
->NodeTypeCode
= NODE_TYPE_PRIVATE_MAP
;
1486 PrivateMap
->ReadAheadMask
= PAGE_SIZE
- 1;
1487 PrivateMap
->FileObject
= FileObject
;
1488 KeInitializeSpinLock(&PrivateMap
->ReadAheadSpinLock
);
1490 /* Link it to the file */
1491 KeAcquireSpinLockAtDpcLevel(&SharedCacheMap
->CacheMapLock
);
1492 InsertTailList(&SharedCacheMap
->PrivateList
, &PrivateMap
->PrivateLinks
);
1493 KeReleaseSpinLockFromDpcLevel(&SharedCacheMap
->CacheMapLock
);
1495 FileObject
->PrivateCacheMap
= PrivateMap
;
1496 SharedCacheMap
->OpenCount
++;
1498 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1500 return STATUS_SUCCESS
;
1508 CcGetFileObjectFromSectionPtrs (
1509 IN PSECTION_OBJECT_POINTERS SectionObjectPointers
)
1511 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1513 CCTRACE(CC_API_DEBUG
, "SectionObjectPointers=%p\n", SectionObjectPointers
);
1515 if (SectionObjectPointers
&& SectionObjectPointers
->SharedCacheMap
)
1517 SharedCacheMap
= SectionObjectPointers
->SharedCacheMap
;
1518 ASSERT(SharedCacheMap
);
1519 return SharedCacheMap
->FileObject
;
1530 DPRINT("CcInitView()\n");
1532 InitializeListHead(&DirtyVacbListHead
);
1533 InitializeListHead(&VacbLruListHead
);
1534 InitializeListHead(&CcDeferredWrites
);
1535 InitializeListHead(&CcCleanSharedCacheMapList
);
1536 KeInitializeSpinLock(&CcDeferredWriteSpinLock
);
1537 ExInitializeNPagedLookasideList(&iBcbLookasideList
,
1541 sizeof(INTERNAL_BCB
),
1544 ExInitializeNPagedLookasideList(&SharedCacheMapLookasideList
,
1548 sizeof(ROS_SHARED_CACHE_MAP
),
1549 TAG_SHARED_CACHE_MAP
,
1551 ExInitializeNPagedLookasideList(&VacbLookasideList
,
1559 MmInitializeMemoryConsumer(MC_CACHE
, CcRosTrimCache
);
1561 CcInitCacheZeroPage();
1564 #if DBG && defined(KDBG)
1566 ExpKdbgExtFileCache(ULONG Argc
, PCHAR Argv
[])
1568 PLIST_ENTRY ListEntry
;
1569 UNICODE_STRING NoName
= RTL_CONSTANT_STRING(L
"No name for File");
1571 KdbpPrint(" Usage Summary (in kb)\n");
1572 KdbpPrint("Shared\t\tValid\tDirty\tName\n");
1573 /* No need to lock the spin lock here, we're in DBG */
1574 for (ListEntry
= CcCleanSharedCacheMapList
.Flink
;
1575 ListEntry
!= &CcCleanSharedCacheMapList
;
1576 ListEntry
= ListEntry
->Flink
)
1579 ULONG Valid
= 0, Dirty
= 0;
1580 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1581 PUNICODE_STRING FileName
;
1584 SharedCacheMap
= CONTAINING_RECORD(ListEntry
, ROS_SHARED_CACHE_MAP
, SharedCacheMapLinks
);
1587 Dirty
= (SharedCacheMap
->DirtyPages
* PAGE_SIZE
) / 1024;
1589 /* First, count for all the associated VACB */
1590 for (Vacbs
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
1591 Vacbs
!= &SharedCacheMap
->CacheMapVacbListHead
;
1592 Vacbs
= Vacbs
->Flink
)
1596 Vacb
= CONTAINING_RECORD(Vacbs
, ROS_VACB
, CacheMapVacbListEntry
);
1599 Valid
+= VACB_MAPPING_GRANULARITY
/ 1024;
1604 if (SharedCacheMap
->FileObject
!= NULL
&&
1605 SharedCacheMap
->FileObject
->FileName
.Length
!= 0)
1607 FileName
= &SharedCacheMap
->FileObject
->FileName
;
1609 else if (SharedCacheMap
->FileObject
!= NULL
&&
1610 SharedCacheMap
->FileObject
->FsContext
!= NULL
&&
1611 ((PFSRTL_COMMON_FCB_HEADER
)(SharedCacheMap
->FileObject
->FsContext
))->NodeTypeCode
== 0x0502 &&
1612 ((PFSRTL_COMMON_FCB_HEADER
)(SharedCacheMap
->FileObject
->FsContext
))->NodeByteSize
== 0x1F8 &&
1613 ((PUNICODE_STRING
)(((PUCHAR
)SharedCacheMap
->FileObject
->FsContext
) + 0x100))->Length
!= 0)
1615 FileName
= (PUNICODE_STRING
)(((PUCHAR
)SharedCacheMap
->FileObject
->FsContext
) + 0x100);
1616 Extra
= L
" (FastFAT)";
1624 KdbpPrint("%p\t%d\t%d\t%wZ%S\n", SharedCacheMap
, Valid
, Dirty
, FileName
, Extra
);
1631 ExpKdbgExtDefWrites(ULONG Argc
, PCHAR Argv
[])
1633 KdbpPrint("CcTotalDirtyPages:\t%lu (%lu Kb)\n", CcTotalDirtyPages
,
1634 (CcTotalDirtyPages
* PAGE_SIZE
) / 1024);
1635 KdbpPrint("CcDirtyPageThreshold:\t%lu (%lu Kb)\n", CcDirtyPageThreshold
,
1636 (CcDirtyPageThreshold
* PAGE_SIZE
) / 1024);
1637 KdbpPrint("MmAvailablePages:\t%lu (%lu Kb)\n", MmAvailablePages
,
1638 (MmAvailablePages
* PAGE_SIZE
) / 1024);
1639 KdbpPrint("MmThrottleTop:\t\t%lu (%lu Kb)\n", MmThrottleTop
,
1640 (MmThrottleTop
* PAGE_SIZE
) / 1024);
1641 KdbpPrint("MmThrottleBottom:\t%lu (%lu Kb)\n", MmThrottleBottom
,
1642 (MmThrottleBottom
* PAGE_SIZE
) / 1024);
1643 KdbpPrint("MmModifiedPageListHead.Total:\t%lu (%lu Kb)\n", MmModifiedPageListHead
.Total
,
1644 (MmModifiedPageListHead
.Total
* PAGE_SIZE
) / 1024);
1646 if (CcTotalDirtyPages
>= CcDirtyPageThreshold
)
1648 KdbpPrint("CcTotalDirtyPages above the threshold, writes should be throttled\n");
1650 else if (CcTotalDirtyPages
+ 64 >= CcDirtyPageThreshold
)
1652 KdbpPrint("CcTotalDirtyPages within 64 (max charge) pages of the threshold, writes may be throttled\n");
1656 KdbpPrint("CcTotalDirtyPages below the threshold, writes should not be throttled\n");