2 * COPYRIGHT: See COPYING in the top level directory
3 * PROJECT: ReactOS kernel
4 * FILE: ntoskrnl/cc/view.c
5 * PURPOSE: Cache manager
7 * PROGRAMMERS: David Welch (welch@mcmail.com)
8 * Pierre Schweitzer (pierre@reactos.org)
11 /* NOTES **********************************************************************
13 * This is not the NT implementation of a file cache nor anything much like
16 * The general procedure for a filesystem to implement a read or write
17 * dispatch routine is as follows
19 * (1) If caching for the FCB hasn't been initiated then so do by calling
20 * CcInitializeFileCache.
22 * (2) For each 4k region which is being read or written obtain a cache page
23 * by calling CcRequestCachePage.
25 * (3) If either the page is being read or not completely written, and it is
26 * not up to date then read its data from the underlying medium. If the read
27 * fails then call CcReleaseCachePage with VALID as FALSE and return a error.
29 * (4) Copy the data into or out of the page as necessary.
31 * (5) Release the cache page
33 /* INCLUDES ******************************************************************/
39 #if defined (ALLOC_PRAGMA)
40 #pragma alloc_text(INIT, CcInitView)
43 /* GLOBALS *******************************************************************/
45 LIST_ENTRY DirtyVacbListHead
;
46 static LIST_ENTRY VacbLruListHead
;
48 KGUARDED_MUTEX ViewLock
;
50 NPAGED_LOOKASIDE_LIST iBcbLookasideList
;
51 static NPAGED_LOOKASIDE_LIST SharedCacheMapLookasideList
;
52 static NPAGED_LOOKASIDE_LIST VacbLookasideList
;
54 /* Internal vars (MS):
55 * - Threshold above which lazy writer will start action
56 * - Amount of dirty pages
57 * - List for deferred writes
58 * - Spinlock when dealing with the deferred list
59 * - List for "clean" shared cache maps
61 ULONG CcDirtyPageThreshold
= 0;
62 ULONG CcTotalDirtyPages
= 0;
63 LIST_ENTRY CcDeferredWrites
;
64 KSPIN_LOCK CcDeferredWriteSpinLock
;
65 LIST_ENTRY CcCleanSharedCacheMapList
;
68 ULONG
CcRosVacbIncRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
72 Refs
= InterlockedIncrement((PLONG
)&vacb
->ReferenceCount
);
73 if (vacb
->SharedCacheMap
->Trace
)
75 DbgPrint("(%s:%i) VACB %p ++RefCount=%lu, Dirty %u, PageOut %lu\n",
76 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
81 ULONG
CcRosVacbDecRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
85 Refs
= InterlockedDecrement((PLONG
)&vacb
->ReferenceCount
);
86 ASSERT(!(Refs
== 0 && vacb
->Dirty
));
87 if (vacb
->SharedCacheMap
->Trace
)
89 DbgPrint("(%s:%i) VACB %p --RefCount=%lu, Dirty %u, PageOut %lu\n",
90 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
95 CcRosInternalFreeVacb(vacb
);
100 ULONG
CcRosVacbGetRefCount_(PROS_VACB vacb
, PCSTR file
, INT line
)
104 Refs
= InterlockedCompareExchange((PLONG
)&vacb
->ReferenceCount
, 0, 0);
105 if (vacb
->SharedCacheMap
->Trace
)
107 DbgPrint("(%s:%i) VACB %p ==RefCount=%lu, Dirty %u, PageOut %lu\n",
108 file
, line
, vacb
, Refs
, vacb
->Dirty
, vacb
->PageOut
);
116 /* FUNCTIONS *****************************************************************/
121 PROS_SHARED_CACHE_MAP SharedCacheMap
,
126 PLIST_ENTRY current_entry
;
132 SharedCacheMap
->Trace
= Trace
;
136 DPRINT1("Enabling Tracing for CacheMap 0x%p:\n", SharedCacheMap
);
138 KeAcquireGuardedMutex(&ViewLock
);
139 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldirql
);
141 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
142 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
144 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
145 current_entry
= current_entry
->Flink
;
147 DPRINT1(" VACB 0x%p enabled, RefCount %lu, Dirty %u, PageOut %lu\n",
148 current
, current
->ReferenceCount
, current
->Dirty
, current
->PageOut
);
150 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldirql
);
151 KeReleaseGuardedMutex(&ViewLock
);
155 DPRINT1("Disabling Tracing for CacheMap 0x%p:\n", SharedCacheMap
);
159 UNREFERENCED_PARAMETER(SharedCacheMap
);
160 UNREFERENCED_PARAMETER(Trace
);
171 CcRosUnmarkDirtyVacb(Vacb
, TRUE
);
173 Status
= CcWriteVirtualAddress(Vacb
);
174 if (!NT_SUCCESS(Status
))
176 CcRosMarkDirtyVacb(Vacb
);
184 CcRosFlushDirtyPages (
188 BOOLEAN CalledFromLazy
)
190 PLIST_ENTRY current_entry
;
195 DPRINT("CcRosFlushDirtyPages(Target %lu)\n", Target
);
199 KeEnterCriticalRegion();
200 KeAcquireGuardedMutex(&ViewLock
);
202 current_entry
= DirtyVacbListHead
.Flink
;
203 if (current_entry
== &DirtyVacbListHead
)
205 DPRINT("No Dirty pages\n");
208 while ((current_entry
!= &DirtyVacbListHead
) && (Target
> 0))
210 current
= CONTAINING_RECORD(current_entry
,
213 current_entry
= current_entry
->Flink
;
215 CcRosVacbIncRefCount(current
);
217 /* When performing lazy write, don't handle temporary files */
218 if (CalledFromLazy
&&
219 BooleanFlagOn(current
->SharedCacheMap
->FileObject
->Flags
, FO_TEMPORARY_FILE
))
221 CcRosVacbDecRefCount(current
);
225 Locked
= current
->SharedCacheMap
->Callbacks
->AcquireForLazyWrite(
226 current
->SharedCacheMap
->LazyWriteContext
, Wait
);
229 CcRosVacbDecRefCount(current
);
233 ASSERT(current
->Dirty
);
235 KeReleaseGuardedMutex(&ViewLock
);
237 Status
= CcRosFlushVacb(current
);
239 current
->SharedCacheMap
->Callbacks
->ReleaseFromLazyWrite(
240 current
->SharedCacheMap
->LazyWriteContext
);
242 KeAcquireGuardedMutex(&ViewLock
);
243 CcRosVacbDecRefCount(current
);
245 if (!NT_SUCCESS(Status
) && (Status
!= STATUS_END_OF_FILE
) &&
246 (Status
!= STATUS_MEDIA_WRITE_PROTECTED
))
248 DPRINT1("CC: Failed to flush VACB.\n");
254 /* How many pages did we free? */
255 PagesFreed
= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
256 (*Count
) += PagesFreed
;
258 /* Make sure we don't overflow target! */
259 if (Target
< PagesFreed
)
261 /* If we would have, jump to zero directly */
266 Target
-= PagesFreed
;
270 current_entry
= DirtyVacbListHead
.Flink
;
273 KeReleaseGuardedMutex(&ViewLock
);
274 KeLeaveCriticalRegion();
276 DPRINT("CcRosFlushDirtyPages() finished\n");
277 return STATUS_SUCCESS
;
286 * FUNCTION: Try to free some memory from the file cache.
288 * Target - The number of pages to be freed.
289 * Priority - The priority of free (currently unused).
290 * NrFreed - Points to a variable where the number of pages
291 * actually freed is returned.
294 PLIST_ENTRY current_entry
;
301 BOOLEAN FlushedPages
= FALSE
;
303 DPRINT("CcRosTrimCache(Target %lu)\n", Target
);
305 InitializeListHead(&FreeList
);
310 KeAcquireGuardedMutex(&ViewLock
);
312 current_entry
= VacbLruListHead
.Flink
;
313 while (current_entry
!= &VacbLruListHead
)
317 current
= CONTAINING_RECORD(current_entry
,
320 current_entry
= current_entry
->Flink
;
322 KeAcquireSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, &oldIrql
);
324 /* Reference the VACB */
325 CcRosVacbIncRefCount(current
);
327 /* Check if it's mapped and not dirty */
328 if (InterlockedCompareExchange((PLONG
)¤t
->MappedCount
, 0, 0) > 0 && !current
->Dirty
)
330 /* We have to break these locks because Cc sucks */
331 KeReleaseSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, oldIrql
);
332 KeReleaseGuardedMutex(&ViewLock
);
334 /* Page out the VACB */
335 for (i
= 0; i
< VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
; i
++)
337 Page
= (PFN_NUMBER
)(MmGetPhysicalAddress((PUCHAR
)current
->BaseAddress
+ (i
* PAGE_SIZE
)).QuadPart
>> PAGE_SHIFT
);
339 MmPageOutPhysicalAddress(Page
);
342 /* Reacquire the locks */
343 KeAcquireGuardedMutex(&ViewLock
);
344 KeAcquireSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, &oldIrql
);
347 /* Dereference the VACB */
348 Refs
= CcRosVacbDecRefCount(current
);
350 /* Check if we can free this entry now */
353 ASSERT(!current
->Dirty
);
354 ASSERT(!current
->MappedCount
);
357 RemoveEntryList(¤t
->CacheMapVacbListEntry
);
358 RemoveEntryList(¤t
->VacbLruListEntry
);
359 InitializeListHead(¤t
->VacbLruListEntry
);
360 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
362 /* Calculate how many pages we freed for Mm */
363 PagesFreed
= min(VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
, Target
);
364 Target
-= PagesFreed
;
365 (*NrFreed
) += PagesFreed
;
368 KeReleaseSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, oldIrql
);
371 KeReleaseGuardedMutex(&ViewLock
);
373 /* Try flushing pages if we haven't met our target */
374 if ((Target
> 0) && !FlushedPages
)
376 /* Flush dirty pages to disk */
377 CcRosFlushDirtyPages(Target
, &PagesFreed
, FALSE
, FALSE
);
380 /* We can only swap as many pages as we flushed */
381 if (PagesFreed
< Target
) Target
= PagesFreed
;
383 /* Check if we flushed anything */
386 /* Try again after flushing dirty pages */
387 DPRINT("Flushed %lu dirty cache pages to disk\n", PagesFreed
);
392 while (!IsListEmpty(&FreeList
))
396 current_entry
= RemoveHeadList(&FreeList
);
397 current
= CONTAINING_RECORD(current_entry
,
399 CacheMapVacbListEntry
);
400 InitializeListHead(¤t
->CacheMapVacbListEntry
);
401 Refs
= CcRosVacbDecRefCount(current
);
405 DPRINT("Evicted %lu cache pages\n", (*NrFreed
));
407 return STATUS_SUCCESS
;
413 PROS_SHARED_CACHE_MAP SharedCacheMap
,
420 ASSERT(SharedCacheMap
);
422 DPRINT("CcRosReleaseVacb(SharedCacheMap 0x%p, Vacb 0x%p, Valid %u)\n",
423 SharedCacheMap
, Vacb
, Valid
);
427 if (Dirty
&& !Vacb
->Dirty
)
429 CcRosMarkDirtyVacb(Vacb
);
434 if (InterlockedIncrement((PLONG
)&Vacb
->MappedCount
) == 1)
436 CcRosVacbIncRefCount(Vacb
);
440 Refs
= CcRosVacbDecRefCount(Vacb
);
443 return STATUS_SUCCESS
;
446 /* Returns with VACB Lock Held! */
450 PROS_SHARED_CACHE_MAP SharedCacheMap
,
453 PLIST_ENTRY current_entry
;
457 ASSERT(SharedCacheMap
);
459 DPRINT("CcRosLookupVacb(SharedCacheMap 0x%p, FileOffset %I64u)\n",
460 SharedCacheMap
, FileOffset
);
462 KeAcquireGuardedMutex(&ViewLock
);
463 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
465 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
466 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
468 current
= CONTAINING_RECORD(current_entry
,
470 CacheMapVacbListEntry
);
471 if (IsPointInRange(current
->FileOffset
.QuadPart
,
472 VACB_MAPPING_GRANULARITY
,
475 CcRosVacbIncRefCount(current
);
476 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
477 KeReleaseGuardedMutex(&ViewLock
);
480 if (current
->FileOffset
.QuadPart
> FileOffset
)
482 current_entry
= current_entry
->Flink
;
485 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
486 KeReleaseGuardedMutex(&ViewLock
);
497 PROS_SHARED_CACHE_MAP SharedCacheMap
;
499 SharedCacheMap
= Vacb
->SharedCacheMap
;
501 KeAcquireGuardedMutex(&ViewLock
);
502 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
504 ASSERT(!Vacb
->Dirty
);
506 InsertTailList(&DirtyVacbListHead
, &Vacb
->DirtyVacbListEntry
);
507 CcTotalDirtyPages
+= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
508 Vacb
->SharedCacheMap
->DirtyPages
+= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
509 CcRosVacbIncRefCount(Vacb
);
511 /* Move to the tail of the LRU list */
512 RemoveEntryList(&Vacb
->VacbLruListEntry
);
513 InsertTailList(&VacbLruListHead
, &Vacb
->VacbLruListEntry
);
517 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
518 KeReleaseGuardedMutex(&ViewLock
);
520 /* Schedule a lazy writer run to now that we have dirty VACB */
521 oldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
522 if (!LazyWriter
.ScanActive
)
524 CcScheduleLazyWriteScan(FALSE
);
526 KeReleaseQueuedSpinLock(LockQueueMasterLock
, oldIrql
);
531 CcRosUnmarkDirtyVacb (
536 PROS_SHARED_CACHE_MAP SharedCacheMap
;
538 SharedCacheMap
= Vacb
->SharedCacheMap
;
542 KeAcquireGuardedMutex(&ViewLock
);
543 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
550 RemoveEntryList(&Vacb
->DirtyVacbListEntry
);
551 InitializeListHead(&Vacb
->DirtyVacbListEntry
);
552 CcTotalDirtyPages
-= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
553 Vacb
->SharedCacheMap
->DirtyPages
-= VACB_MAPPING_GRANULARITY
/ PAGE_SIZE
;
554 CcRosVacbDecRefCount(Vacb
);
558 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
559 KeReleaseGuardedMutex(&ViewLock
);
566 PROS_SHARED_CACHE_MAP SharedCacheMap
,
571 ASSERT(SharedCacheMap
);
573 DPRINT("CcRosMarkDirtyVacb(SharedCacheMap 0x%p, FileOffset %I64u)\n",
574 SharedCacheMap
, FileOffset
);
576 Vacb
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
579 KeBugCheck(CACHE_MANAGER
);
582 CcRosReleaseVacb(SharedCacheMap
, Vacb
, Vacb
->Valid
, TRUE
, FALSE
);
584 return STATUS_SUCCESS
;
588 * Note: this is not the contrary function of
589 * CcRosMapVacbInKernelSpace()
594 PROS_SHARED_CACHE_MAP SharedCacheMap
,
600 ASSERT(SharedCacheMap
);
602 DPRINT("CcRosUnmapVacb(SharedCacheMap 0x%p, FileOffset %I64u, NowDirty %u)\n",
603 SharedCacheMap
, FileOffset
, NowDirty
);
605 Vacb
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
608 return STATUS_UNSUCCESSFUL
;
611 ASSERT(Vacb
->MappedCount
!= 0);
612 if (InterlockedDecrement((PLONG
)&Vacb
->MappedCount
) == 0)
614 CcRosVacbDecRefCount(Vacb
);
617 CcRosReleaseVacb(SharedCacheMap
, Vacb
, Vacb
->Valid
, NowDirty
, FALSE
);
619 return STATUS_SUCCESS
;
624 CcRosMapVacbInKernelSpace(
629 ULONG_PTR NumberOfPages
;
630 PVOID BaseAddress
= NULL
;
632 /* Create a memory area. */
633 MmLockAddressSpace(MmGetKernelAddressSpace());
634 Status
= MmCreateMemoryArea(MmGetKernelAddressSpace(),
635 0, // nothing checks for VACB mareas, so set to 0
637 VACB_MAPPING_GRANULARITY
,
639 (PMEMORY_AREA
*)&Vacb
->MemoryArea
,
642 ASSERT(Vacb
->BaseAddress
== NULL
);
643 Vacb
->BaseAddress
= BaseAddress
;
644 MmUnlockAddressSpace(MmGetKernelAddressSpace());
645 if (!NT_SUCCESS(Status
))
647 DPRINT1("MmCreateMemoryArea failed with %lx for VACB %p\n", Status
, Vacb
);
651 ASSERT(((ULONG_PTR
)Vacb
->BaseAddress
% PAGE_SIZE
) == 0);
652 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
> (ULONG_PTR
)MmSystemRangeStart
);
653 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ VACB_MAPPING_GRANULARITY
- 1 > (ULONG_PTR
)MmSystemRangeStart
);
655 /* Create a virtual mapping for this memory area */
656 NumberOfPages
= BYTES_TO_PAGES(VACB_MAPPING_GRANULARITY
);
657 for (i
= 0; i
< NumberOfPages
; i
++)
659 PFN_NUMBER PageFrameNumber
;
661 MI_SET_USAGE(MI_USAGE_CACHE
);
662 Status
= MmRequestPageMemoryConsumer(MC_CACHE
, TRUE
, &PageFrameNumber
);
663 if (PageFrameNumber
== 0)
665 DPRINT1("Unable to allocate page\n");
666 KeBugCheck(MEMORY_MANAGEMENT
);
669 ASSERT(BaseAddress
== Vacb
->BaseAddress
);
670 ASSERT(i
* PAGE_SIZE
< VACB_MAPPING_GRANULARITY
);
671 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
) >= (ULONG_PTR
)BaseAddress
);
672 ASSERT((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
) > (ULONG_PTR
)MmSystemRangeStart
);
674 Status
= MmCreateVirtualMapping(NULL
,
675 (PVOID
)((ULONG_PTR
)Vacb
->BaseAddress
+ (i
* PAGE_SIZE
)),
679 if (!NT_SUCCESS(Status
))
681 DPRINT1("Unable to create virtual mapping\n");
682 KeBugCheck(MEMORY_MANAGEMENT
);
686 return STATUS_SUCCESS
;
691 CcRosFreeUnusedVacb (
699 PLIST_ENTRY current_entry
;
703 InitializeListHead(&FreeList
);
705 KeAcquireGuardedMutex(&ViewLock
);
707 /* Browse all the available VACB */
708 current_entry
= VacbLruListHead
.Flink
;
709 while (current_entry
!= &VacbLruListHead
)
713 current
= CONTAINING_RECORD(current_entry
,
716 current_entry
= current_entry
->Flink
;
718 KeAcquireSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, &oldIrql
);
720 /* Only deal with unused VACB, we will free them */
721 Refs
= CcRosVacbGetRefCount(current
);
724 ASSERT(!current
->Dirty
);
725 ASSERT(!current
->MappedCount
);
728 /* Reset and move to free list */
729 RemoveEntryList(¤t
->CacheMapVacbListEntry
);
730 RemoveEntryList(¤t
->VacbLruListEntry
);
731 InitializeListHead(¤t
->VacbLruListEntry
);
732 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
735 KeReleaseSpinLock(¤t
->SharedCacheMap
->CacheMapLock
, oldIrql
);
739 KeReleaseGuardedMutex(&ViewLock
);
741 /* And now, free any of the found VACB, that'll free memory! */
742 while (!IsListEmpty(&FreeList
))
746 current_entry
= RemoveHeadList(&FreeList
);
747 current
= CONTAINING_RECORD(current_entry
,
749 CacheMapVacbListEntry
);
750 InitializeListHead(¤t
->CacheMapVacbListEntry
);
751 Refs
= CcRosVacbDecRefCount(current
);
756 /* If we freed at least one VACB, return success */
762 /* If caller asked for free count, return it */
774 PROS_SHARED_CACHE_MAP SharedCacheMap
,
780 PLIST_ENTRY current_entry
;
786 ASSERT(SharedCacheMap
);
788 DPRINT("CcRosCreateVacb()\n");
790 if (FileOffset
>= SharedCacheMap
->SectionSize
.QuadPart
)
793 return STATUS_INVALID_PARAMETER
;
796 current
= ExAllocateFromNPagedLookasideList(&VacbLookasideList
);
797 current
->BaseAddress
= NULL
;
798 current
->Valid
= FALSE
;
799 current
->Dirty
= FALSE
;
800 current
->PageOut
= FALSE
;
801 current
->FileOffset
.QuadPart
= ROUND_DOWN(FileOffset
, VACB_MAPPING_GRANULARITY
);
802 current
->SharedCacheMap
= SharedCacheMap
;
804 if (SharedCacheMap
->Trace
)
806 DPRINT1("CacheMap 0x%p: new VACB: 0x%p\n", SharedCacheMap
, current
);
809 current
->MappedCount
= 0;
810 current
->ReferenceCount
= 0;
811 InitializeListHead(¤t
->CacheMapVacbListEntry
);
812 InitializeListHead(¤t
->DirtyVacbListEntry
);
813 InitializeListHead(¤t
->VacbLruListEntry
);
815 CcRosVacbIncRefCount(current
);
819 /* Map VACB in kernel space */
820 Status
= CcRosMapVacbInKernelSpace(current
);
821 if (!NT_SUCCESS(Status
))
824 /* If no space left, try to prune unused VACB
825 * to recover space to map our VACB
826 * If it succeed, retry to map, otherwise
829 if (!Retried
&& CcRosFreeUnusedVacb(&Freed
))
831 DPRINT("Prunned %d VACB, trying again\n", Freed
);
836 CcRosVacbDecRefCount(current
);
837 ExFreeToNPagedLookasideList(&VacbLookasideList
, current
);
841 KeAcquireGuardedMutex(&ViewLock
);
844 /* There is window between the call to CcRosLookupVacb
845 * and CcRosCreateVacb. We must check if a VACB for the
846 * file offset exist. If there is a VACB, we release
847 * our newly created VACB and return the existing one.
849 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
850 current_entry
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
852 while (current_entry
!= &SharedCacheMap
->CacheMapVacbListHead
)
854 current
= CONTAINING_RECORD(current_entry
,
856 CacheMapVacbListEntry
);
857 if (IsPointInRange(current
->FileOffset
.QuadPart
,
858 VACB_MAPPING_GRANULARITY
,
861 CcRosVacbIncRefCount(current
);
862 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
864 if (SharedCacheMap
->Trace
)
866 DPRINT1("CacheMap 0x%p: deleting newly created VACB 0x%p ( found existing one 0x%p )\n",
872 KeReleaseGuardedMutex(&ViewLock
);
874 Refs
= CcRosVacbDecRefCount(*Vacb
);
878 return STATUS_SUCCESS
;
880 if (current
->FileOffset
.QuadPart
< FileOffset
)
882 ASSERT(previous
== NULL
||
883 previous
->FileOffset
.QuadPart
< current
->FileOffset
.QuadPart
);
886 if (current
->FileOffset
.QuadPart
> FileOffset
)
888 current_entry
= current_entry
->Flink
;
890 /* There was no existing VACB. */
894 InsertHeadList(&previous
->CacheMapVacbListEntry
, ¤t
->CacheMapVacbListEntry
);
898 InsertHeadList(&SharedCacheMap
->CacheMapVacbListHead
, ¤t
->CacheMapVacbListEntry
);
900 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
901 InsertTailList(&VacbLruListHead
, ¤t
->VacbLruListEntry
);
902 KeReleaseGuardedMutex(&ViewLock
);
904 MI_SET_USAGE(MI_USAGE_CACHE
);
906 if ((SharedCacheMap
->FileObject
) && (SharedCacheMap
->FileObject
->FileName
.Buffer
))
910 pos
= wcsrchr(SharedCacheMap
->FileObject
->FileName
.Buffer
, '\\');
913 len
= wcslen(pos
) * sizeof(WCHAR
);
914 snprintf(MI_PFN_CURRENT_PROCESS_NAME
, min(16, len
), "%S", pos
);
918 snprintf(MI_PFN_CURRENT_PROCESS_NAME
, min(16, len
), "%wZ", &SharedCacheMap
->FileObject
->FileName
);
923 /* Reference it to allow release */
924 CcRosVacbIncRefCount(current
);
932 PROS_SHARED_CACHE_MAP SharedCacheMap
,
934 PLONGLONG BaseOffset
,
943 ASSERT(SharedCacheMap
);
945 DPRINT("CcRosGetVacb()\n");
948 * Look for a VACB already mapping the same data.
950 current
= CcRosLookupVacb(SharedCacheMap
, FileOffset
);
954 * Otherwise create a new VACB.
956 Status
= CcRosCreateVacb(SharedCacheMap
, FileOffset
, ¤t
);
957 if (!NT_SUCCESS(Status
))
963 Refs
= CcRosVacbGetRefCount(current
);
965 KeAcquireGuardedMutex(&ViewLock
);
967 /* Move to the tail of the LRU list */
968 RemoveEntryList(¤t
->VacbLruListEntry
);
969 InsertTailList(&VacbLruListHead
, ¤t
->VacbLruListEntry
);
971 KeReleaseGuardedMutex(&ViewLock
);
974 * Return information about the VACB to the caller.
976 *UptoDate
= current
->Valid
;
977 *BaseAddress
= current
->BaseAddress
;
978 DPRINT("*BaseAddress %p\n", *BaseAddress
);
980 *BaseOffset
= current
->FileOffset
.QuadPart
;
984 return STATUS_SUCCESS
;
990 PROS_SHARED_CACHE_MAP SharedCacheMap
,
996 * FUNCTION: Request a page mapping for a shared cache map
1001 ASSERT(SharedCacheMap
);
1003 if (FileOffset
% VACB_MAPPING_GRANULARITY
!= 0)
1005 DPRINT1("Bad fileoffset %I64x should be multiple of %x",
1006 FileOffset
, VACB_MAPPING_GRANULARITY
);
1007 KeBugCheck(CACHE_MANAGER
);
1010 return CcRosGetVacb(SharedCacheMap
,
1022 MEMORY_AREA
* MemoryArea
,
1025 SWAPENTRY SwapEntry
,
1028 ASSERT(SwapEntry
== 0);
1031 ASSERT(MmGetReferenceCountPage(Page
) == 1);
1032 MmReleasePageMemoryConsumer(MC_CACHE
, Page
);
1037 CcRosInternalFreeVacb (
1040 * FUNCTION: Releases a VACB associated with a shared cache map
1043 DPRINT("Freeing VACB 0x%p\n", Vacb
);
1045 if (Vacb
->SharedCacheMap
->Trace
)
1047 DPRINT1("CacheMap 0x%p: deleting VACB: 0x%p\n", Vacb
->SharedCacheMap
, Vacb
);
1051 MmLockAddressSpace(MmGetKernelAddressSpace());
1052 MmFreeMemoryArea(MmGetKernelAddressSpace(),
1056 MmUnlockAddressSpace(MmGetKernelAddressSpace());
1058 if (Vacb
->ReferenceCount
!= 0)
1060 DPRINT1("Invalid free: %ld\n", Vacb
->ReferenceCount
);
1061 if (Vacb
->SharedCacheMap
->FileObject
&& Vacb
->SharedCacheMap
->FileObject
->FileName
.Length
)
1063 DPRINT1("For file: %wZ\n", &Vacb
->SharedCacheMap
->FileObject
->FileName
);
1067 ASSERT(Vacb
->ReferenceCount
== 0);
1068 ASSERT(IsListEmpty(&Vacb
->CacheMapVacbListEntry
));
1069 ASSERT(IsListEmpty(&Vacb
->DirtyVacbListEntry
));
1070 ASSERT(IsListEmpty(&Vacb
->VacbLruListEntry
));
1071 RtlFillMemory(Vacb
, sizeof(*Vacb
), 0xfd);
1072 ExFreeToNPagedLookasideList(&VacbLookasideList
, Vacb
);
1073 return STATUS_SUCCESS
;
1082 IN PSECTION_OBJECT_POINTERS SectionObjectPointers
,
1083 IN PLARGE_INTEGER FileOffset OPTIONAL
,
1085 OUT PIO_STATUS_BLOCK IoStatus
)
1087 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1088 LARGE_INTEGER Offset
;
1089 LONGLONG RemainingLength
;
1093 CCTRACE(CC_API_DEBUG
, "SectionObjectPointers=%p FileOffset=%p Length=%lu\n",
1094 SectionObjectPointers
, FileOffset
, Length
);
1096 DPRINT("CcFlushCache(SectionObjectPointers 0x%p, FileOffset 0x%p, Length %lu, IoStatus 0x%p)\n",
1097 SectionObjectPointers
, FileOffset
, Length
, IoStatus
);
1099 if (SectionObjectPointers
&& SectionObjectPointers
->SharedCacheMap
)
1101 SharedCacheMap
= SectionObjectPointers
->SharedCacheMap
;
1102 ASSERT(SharedCacheMap
);
1105 Offset
= *FileOffset
;
1106 RemainingLength
= Length
;
1110 Offset
.QuadPart
= 0;
1111 RemainingLength
= SharedCacheMap
->FileSize
.QuadPart
;
1116 IoStatus
->Status
= STATUS_SUCCESS
;
1117 IoStatus
->Information
= 0;
1120 while (RemainingLength
> 0)
1122 current
= CcRosLookupVacb(SharedCacheMap
, Offset
.QuadPart
);
1123 if (current
!= NULL
)
1127 Status
= CcRosFlushVacb(current
);
1128 if (!NT_SUCCESS(Status
) && IoStatus
!= NULL
)
1130 IoStatus
->Status
= Status
;
1134 CcRosReleaseVacb(SharedCacheMap
, current
, current
->Valid
, current
->Dirty
, FALSE
);
1137 Offset
.QuadPart
+= VACB_MAPPING_GRANULARITY
;
1138 RemainingLength
-= min(RemainingLength
, VACB_MAPPING_GRANULARITY
);
1145 IoStatus
->Status
= STATUS_INVALID_PARAMETER
;
1152 CcRosDeleteFileCache (
1153 PFILE_OBJECT FileObject
,
1154 PROS_SHARED_CACHE_MAP SharedCacheMap
)
1156 * FUNCTION: Releases the shared cache map associated with a file object
1159 PLIST_ENTRY current_entry
;
1161 LIST_ENTRY FreeList
;
1164 ASSERT(SharedCacheMap
);
1166 SharedCacheMap
->OpenCount
++;
1167 KeReleaseGuardedMutex(&ViewLock
);
1169 CcFlushCache(FileObject
->SectionObjectPointer
, NULL
, 0, NULL
);
1171 KeAcquireGuardedMutex(&ViewLock
);
1172 SharedCacheMap
->OpenCount
--;
1173 if (SharedCacheMap
->OpenCount
== 0)
1177 FileObject
->SectionObjectPointer
->SharedCacheMap
= NULL
;
1182 InitializeListHead(&FreeList
);
1183 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
1184 while (!IsListEmpty(&SharedCacheMap
->CacheMapVacbListHead
))
1186 current_entry
= RemoveTailList(&SharedCacheMap
->CacheMapVacbListHead
);
1187 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
1189 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
1190 RemoveEntryList(¤t
->VacbLruListEntry
);
1191 InitializeListHead(¤t
->VacbLruListEntry
);
1194 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
1195 CcRosUnmarkDirtyVacb(current
, FALSE
);
1196 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
1197 DPRINT1("Freeing dirty VACB\n");
1199 if (current
->MappedCount
!= 0)
1201 current
->MappedCount
= 0;
1202 NT_VERIFY(CcRosVacbDecRefCount(current
) > 0);
1203 DPRINT1("Freeing mapped VACB\n");
1205 InsertHeadList(&FreeList
, ¤t
->CacheMapVacbListEntry
);
1207 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &oldIrql
);
1210 SharedCacheMap
->Trace
= FALSE
;
1212 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, oldIrql
);
1214 KeReleaseGuardedMutex(&ViewLock
);
1215 ObDereferenceObject(SharedCacheMap
->FileObject
);
1217 while (!IsListEmpty(&FreeList
))
1221 current_entry
= RemoveTailList(&FreeList
);
1222 current
= CONTAINING_RECORD(current_entry
, ROS_VACB
, CacheMapVacbListEntry
);
1223 InitializeListHead(¤t
->CacheMapVacbListEntry
);
1224 Refs
= CcRosVacbDecRefCount(current
);
1225 #if DBG // CORE-14578
1228 DPRINT1("Leaking VACB %p attached to %p (%I64d)\n", current
, FileObject
, current
->FileOffset
.QuadPart
);
1229 DPRINT1("There are: %d references left\n", Refs
);
1230 DPRINT1("Map: %d\n", current
->MappedCount
);
1231 DPRINT1("Dirty: %d\n", current
->Dirty
);
1232 if (FileObject
->FileName
.Length
!= 0)
1234 DPRINT1("File was: %wZ\n", &FileObject
->FileName
);
1236 else if (FileObject
->FsContext
!= NULL
&&
1237 ((PFSRTL_COMMON_FCB_HEADER
)(FileObject
->FsContext
))->NodeTypeCode
== 0x0502 &&
1238 ((PFSRTL_COMMON_FCB_HEADER
)(FileObject
->FsContext
))->NodeByteSize
== 0x1F8 &&
1239 ((PUNICODE_STRING
)(((PUCHAR
)FileObject
->FsContext
) + 0x100))->Length
!= 0)
1241 DPRINT1("File was: %wZ (FastFAT)\n", (PUNICODE_STRING
)(((PUCHAR
)FileObject
->FsContext
) + 0x100));
1245 DPRINT1("No name for the file\n");
1253 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1254 RemoveEntryList(&SharedCacheMap
->SharedCacheMapLinks
);
1255 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1257 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1258 KeAcquireGuardedMutex(&ViewLock
);
1260 return STATUS_SUCCESS
;
1265 CcRosReferenceCache (
1266 PFILE_OBJECT FileObject
)
1268 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1269 KeAcquireGuardedMutex(&ViewLock
);
1270 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1271 ASSERT(SharedCacheMap
);
1272 ASSERT(SharedCacheMap
->OpenCount
!= 0);
1273 SharedCacheMap
->OpenCount
++;
1274 KeReleaseGuardedMutex(&ViewLock
);
1279 CcRosRemoveIfClosed (
1280 PSECTION_OBJECT_POINTERS SectionObjectPointer
)
1282 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1283 DPRINT("CcRosRemoveIfClosed()\n");
1284 KeAcquireGuardedMutex(&ViewLock
);
1285 SharedCacheMap
= SectionObjectPointer
->SharedCacheMap
;
1286 if (SharedCacheMap
&& SharedCacheMap
->OpenCount
== 0)
1288 CcRosDeleteFileCache(SharedCacheMap
->FileObject
, SharedCacheMap
);
1290 KeReleaseGuardedMutex(&ViewLock
);
1296 CcRosDereferenceCache (
1297 PFILE_OBJECT FileObject
)
1299 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1300 KeAcquireGuardedMutex(&ViewLock
);
1301 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1302 ASSERT(SharedCacheMap
);
1303 if (SharedCacheMap
->OpenCount
> 0)
1305 SharedCacheMap
->OpenCount
--;
1306 if (SharedCacheMap
->OpenCount
== 0)
1308 MmFreeSectionSegments(SharedCacheMap
->FileObject
);
1309 CcRosDeleteFileCache(FileObject
, SharedCacheMap
);
1312 KeReleaseGuardedMutex(&ViewLock
);
1317 CcRosReleaseFileCache (
1318 PFILE_OBJECT FileObject
)
1320 * FUNCTION: Called by the file system when a handle to a file object
1325 PPRIVATE_CACHE_MAP PrivateMap
;
1326 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1328 KeAcquireGuardedMutex(&ViewLock
);
1330 if (FileObject
->SectionObjectPointer
->SharedCacheMap
!= NULL
)
1332 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1334 /* Closing the handle, so kill the private cache map
1335 * Before you event try to remove it from FO, always
1336 * lock the master lock, to be sure not to race
1337 * with a potential read ahead ongoing!
1339 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1340 PrivateMap
= FileObject
->PrivateCacheMap
;
1341 FileObject
->PrivateCacheMap
= NULL
;
1342 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1344 if (PrivateMap
!= NULL
)
1346 /* Remove it from the file */
1347 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &OldIrql
);
1348 RemoveEntryList(&PrivateMap
->PrivateLinks
);
1349 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, OldIrql
);
1352 if (PrivateMap
!= &SharedCacheMap
->PrivateCacheMap
)
1354 ExFreePoolWithTag(PrivateMap
, TAG_PRIVATE_CACHE_MAP
);
1358 PrivateMap
->NodeTypeCode
= 0;
1361 if (SharedCacheMap
->OpenCount
> 0)
1363 SharedCacheMap
->OpenCount
--;
1364 if (SharedCacheMap
->OpenCount
== 0)
1366 MmFreeSectionSegments(SharedCacheMap
->FileObject
);
1367 CcRosDeleteFileCache(FileObject
, SharedCacheMap
);
1372 KeReleaseGuardedMutex(&ViewLock
);
1373 return STATUS_SUCCESS
;
1378 CcRosInitializeFileCache (
1379 PFILE_OBJECT FileObject
,
1380 PCC_FILE_SIZES FileSizes
,
1382 PCACHE_MANAGER_CALLBACKS CallBacks
,
1383 PVOID LazyWriterContext
)
1385 * FUNCTION: Initializes a shared cache map for a file object
1390 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1392 DPRINT("CcRosInitializeFileCache(FileObject 0x%p)\n", FileObject
);
1395 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1396 if (SharedCacheMap
== NULL
)
1399 SharedCacheMap
= ExAllocateFromNPagedLookasideList(&SharedCacheMapLookasideList
);
1400 if (SharedCacheMap
== NULL
)
1402 return STATUS_INSUFFICIENT_RESOURCES
;
1404 RtlZeroMemory(SharedCacheMap
, sizeof(*SharedCacheMap
));
1405 SharedCacheMap
->NodeTypeCode
= NODE_TYPE_SHARED_MAP
;
1406 SharedCacheMap
->NodeByteSize
= sizeof(*SharedCacheMap
);
1407 SharedCacheMap
->FileObject
= FileObject
;
1408 SharedCacheMap
->Callbacks
= CallBacks
;
1409 SharedCacheMap
->LazyWriteContext
= LazyWriterContext
;
1410 SharedCacheMap
->SectionSize
= FileSizes
->AllocationSize
;
1411 SharedCacheMap
->FileSize
= FileSizes
->FileSize
;
1412 SharedCacheMap
->PinAccess
= PinAccess
;
1413 SharedCacheMap
->DirtyPageThreshold
= 0;
1414 SharedCacheMap
->DirtyPages
= 0;
1415 InitializeListHead(&SharedCacheMap
->PrivateList
);
1416 KeInitializeSpinLock(&SharedCacheMap
->CacheMapLock
);
1417 InitializeListHead(&SharedCacheMap
->CacheMapVacbListHead
);
1418 InitializeListHead(&SharedCacheMap
->BcbList
);
1421 KeAcquireGuardedMutex(&ViewLock
);
1424 if (FileObject
->SectionObjectPointer
->SharedCacheMap
== NULL
)
1426 ObReferenceObjectByPointer(FileObject
,
1430 FileObject
->SectionObjectPointer
->SharedCacheMap
= SharedCacheMap
;
1432 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1433 InsertTailList(&CcCleanSharedCacheMapList
, &SharedCacheMap
->SharedCacheMapLinks
);
1434 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1438 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1439 SharedCacheMap
= FileObject
->SectionObjectPointer
->SharedCacheMap
;
1442 if (FileObject
->PrivateCacheMap
== NULL
)
1444 PPRIVATE_CACHE_MAP PrivateMap
;
1446 /* Allocate the private cache map for this handle */
1447 if (SharedCacheMap
->PrivateCacheMap
.NodeTypeCode
!= 0)
1449 PrivateMap
= ExAllocatePoolWithTag(NonPagedPool
, sizeof(PRIVATE_CACHE_MAP
), TAG_PRIVATE_CACHE_MAP
);
1453 PrivateMap
= &SharedCacheMap
->PrivateCacheMap
;
1456 if (PrivateMap
== NULL
)
1458 /* If we also allocated the shared cache map for this file, kill it */
1461 OldIrql
= KeAcquireQueuedSpinLock(LockQueueMasterLock
);
1462 RemoveEntryList(&SharedCacheMap
->SharedCacheMapLinks
);
1463 KeReleaseQueuedSpinLock(LockQueueMasterLock
, OldIrql
);
1465 FileObject
->SectionObjectPointer
->SharedCacheMap
= NULL
;
1466 ObDereferenceObject(FileObject
);
1467 ExFreeToNPagedLookasideList(&SharedCacheMapLookasideList
, SharedCacheMap
);
1470 KeReleaseGuardedMutex(&ViewLock
);
1471 return STATUS_INSUFFICIENT_RESOURCES
;
1475 RtlZeroMemory(PrivateMap
, sizeof(PRIVATE_CACHE_MAP
));
1476 PrivateMap
->NodeTypeCode
= NODE_TYPE_PRIVATE_MAP
;
1477 PrivateMap
->ReadAheadMask
= PAGE_SIZE
- 1;
1478 PrivateMap
->FileObject
= FileObject
;
1479 KeInitializeSpinLock(&PrivateMap
->ReadAheadSpinLock
);
1481 /* Link it to the file */
1482 KeAcquireSpinLock(&SharedCacheMap
->CacheMapLock
, &OldIrql
);
1483 InsertTailList(&SharedCacheMap
->PrivateList
, &PrivateMap
->PrivateLinks
);
1484 KeReleaseSpinLock(&SharedCacheMap
->CacheMapLock
, OldIrql
);
1486 FileObject
->PrivateCacheMap
= PrivateMap
;
1487 SharedCacheMap
->OpenCount
++;
1489 KeReleaseGuardedMutex(&ViewLock
);
1491 return STATUS_SUCCESS
;
1499 CcGetFileObjectFromSectionPtrs (
1500 IN PSECTION_OBJECT_POINTERS SectionObjectPointers
)
1502 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1504 CCTRACE(CC_API_DEBUG
, "SectionObjectPointers=%p\n", SectionObjectPointers
);
1506 if (SectionObjectPointers
&& SectionObjectPointers
->SharedCacheMap
)
1508 SharedCacheMap
= SectionObjectPointers
->SharedCacheMap
;
1509 ASSERT(SharedCacheMap
);
1510 return SharedCacheMap
->FileObject
;
1521 DPRINT("CcInitView()\n");
1523 InitializeListHead(&DirtyVacbListHead
);
1524 InitializeListHead(&VacbLruListHead
);
1525 InitializeListHead(&CcDeferredWrites
);
1526 InitializeListHead(&CcCleanSharedCacheMapList
);
1527 KeInitializeSpinLock(&CcDeferredWriteSpinLock
);
1528 KeInitializeGuardedMutex(&ViewLock
);
1529 ExInitializeNPagedLookasideList(&iBcbLookasideList
,
1533 sizeof(INTERNAL_BCB
),
1536 ExInitializeNPagedLookasideList(&SharedCacheMapLookasideList
,
1540 sizeof(ROS_SHARED_CACHE_MAP
),
1541 TAG_SHARED_CACHE_MAP
,
1543 ExInitializeNPagedLookasideList(&VacbLookasideList
,
1551 MmInitializeMemoryConsumer(MC_CACHE
, CcRosTrimCache
);
1553 CcInitCacheZeroPage();
1556 #if DBG && defined(KDBG)
1558 ExpKdbgExtFileCache(ULONG Argc
, PCHAR Argv
[])
1560 PLIST_ENTRY ListEntry
;
1561 UNICODE_STRING NoName
= RTL_CONSTANT_STRING(L
"No name for File");
1563 KdbpPrint(" Usage Summary (in kb)\n");
1564 KdbpPrint("Shared\t\tValid\tDirty\tName\n");
1565 /* No need to lock the spin lock here, we're in DBG */
1566 for (ListEntry
= CcCleanSharedCacheMapList
.Flink
;
1567 ListEntry
!= &CcCleanSharedCacheMapList
;
1568 ListEntry
= ListEntry
->Flink
)
1571 ULONG Valid
= 0, Dirty
= 0;
1572 PROS_SHARED_CACHE_MAP SharedCacheMap
;
1573 PUNICODE_STRING FileName
;
1576 SharedCacheMap
= CONTAINING_RECORD(ListEntry
, ROS_SHARED_CACHE_MAP
, SharedCacheMapLinks
);
1579 Dirty
= (SharedCacheMap
->DirtyPages
* PAGE_SIZE
) / 1024;
1581 /* First, count for all the associated VACB */
1582 for (Vacbs
= SharedCacheMap
->CacheMapVacbListHead
.Flink
;
1583 Vacbs
!= &SharedCacheMap
->CacheMapVacbListHead
;
1584 Vacbs
= Vacbs
->Flink
)
1588 Vacb
= CONTAINING_RECORD(Vacbs
, ROS_VACB
, CacheMapVacbListEntry
);
1591 Valid
+= VACB_MAPPING_GRANULARITY
/ 1024;
1596 if (SharedCacheMap
->FileObject
!= NULL
&&
1597 SharedCacheMap
->FileObject
->FileName
.Length
!= 0)
1599 FileName
= &SharedCacheMap
->FileObject
->FileName
;
1601 else if (SharedCacheMap
->FileObject
!= NULL
&&
1602 SharedCacheMap
->FileObject
->FsContext
!= NULL
&&
1603 ((PFSRTL_COMMON_FCB_HEADER
)(SharedCacheMap
->FileObject
->FsContext
))->NodeTypeCode
== 0x0502 &&
1604 ((PFSRTL_COMMON_FCB_HEADER
)(SharedCacheMap
->FileObject
->FsContext
))->NodeByteSize
== 0x1F8 &&
1605 ((PUNICODE_STRING
)(((PUCHAR
)SharedCacheMap
->FileObject
->FsContext
) + 0x100))->Length
!= 0)
1607 FileName
= (PUNICODE_STRING
)(((PUCHAR
)SharedCacheMap
->FileObject
->FsContext
) + 0x100);
1608 Extra
= L
" (FastFAT)";
1616 KdbpPrint("%p\t%d\t%d\t%wZ%S\n", SharedCacheMap
, Valid
, Dirty
, FileName
, Extra
);
1623 ExpKdbgExtDefWrites(ULONG Argc
, PCHAR Argv
[])
1625 KdbpPrint("CcTotalDirtyPages:\t%lu (%lu Kb)\n", CcTotalDirtyPages
,
1626 (CcTotalDirtyPages
* PAGE_SIZE
) / 1024);
1627 KdbpPrint("CcDirtyPageThreshold:\t%lu (%lu Kb)\n", CcDirtyPageThreshold
,
1628 (CcDirtyPageThreshold
* PAGE_SIZE
) / 1024);
1629 KdbpPrint("MmAvailablePages:\t%lu (%lu Kb)\n", MmAvailablePages
,
1630 (MmAvailablePages
* PAGE_SIZE
) / 1024);
1631 KdbpPrint("MmThrottleTop:\t\t%lu (%lu Kb)\n", MmThrottleTop
,
1632 (MmThrottleTop
* PAGE_SIZE
) / 1024);
1633 KdbpPrint("MmThrottleBottom:\t%lu (%lu Kb)\n", MmThrottleBottom
,
1634 (MmThrottleBottom
* PAGE_SIZE
) / 1024);
1635 KdbpPrint("MmModifiedPageListHead.Total:\t%lu (%lu Kb)\n", MmModifiedPageListHead
.Total
,
1636 (MmModifiedPageListHead
.Total
* PAGE_SIZE
) / 1024);
1638 if (CcTotalDirtyPages
>= CcDirtyPageThreshold
)
1640 KdbpPrint("CcTotalDirtyPages above the threshold, writes should be throttled\n");
1642 else if (CcTotalDirtyPages
+ 64 >= CcDirtyPageThreshold
)
1644 KdbpPrint("CcTotalDirtyPages within 64 (max charge) pages of the threshold, writes may be throttled\n");
1648 KdbpPrint("CcTotalDirtyPages below the threshold, writes should not be throttled\n");