Kernel/x86_64 - Separated task switching from timer interrupt
[tpg/acess2.git] / Kernel / arch / x86_64 / mm_virt.c
index b13c47c..2c91aed 100644 (file)
@@ -6,10 +6,12 @@
 #define DEBUG  0
 #include <acess.h>
 #include <mm_virt.h>
+#include <threads_int.h>
 #include <proc.h>
 
 // === CONSTANTS ===
 #define PHYS_BITS      52      // TODO: Move out
+#define VIRT_BITS      48
 
 #define PML4_SHIFT     39
 #define PDP_SHIFT      30
 #define PTAB_SHIFT     12
 
 #define        PADDR_MASK      0x7FFFFFFF##FFFFF000
-#define PAGE_MASK      (((Uint)1 << 36)-1)
-#define TABLE_MASK     (((Uint)1 << 27)-1)
-#define PDP_MASK       (((Uint)1 << 18)-1)
-#define PML4_MASK      (((Uint)1 << 9)-1)
+#define PAGE_MASK      ((1LL << 36)-1)
+#define TABLE_MASK     ((1LL << 27)-1)
+#define PDP_MASK       ((1LL << 18)-1)
+#define PML4_MASK      ((1LL << 9)-1)
 
 #define        PF_PRESENT      0x001
 #define        PF_WRITE        0x002
 #define        PF_USER         0x004
-#define        PF_LARGE        0x000
+#define        PF_LARGE        0x080
+#define        PF_GLOBAL       0x100
 #define        PF_COW          0x200
 #define        PF_PAGED        0x400
 #define        PF_NX           0x80000000##00000000
 
 // === MACROS ===
-#define PAGETABLE(idx) (*((tPAddr*)MM_FRACTAL_BASE+((idx)&PAGE_MASK)))
+#define PAGETABLE(idx) (*((Uint64*)MM_FRACTAL_BASE+((idx)&PAGE_MASK)))
 #define PAGEDIR(idx)   PAGETABLE((MM_FRACTAL_BASE>>12)+((idx)&TABLE_MASK))
 #define PAGEDIRPTR(idx)        PAGEDIR((MM_FRACTAL_BASE>>21)+((idx)&PDP_MASK))
 #define PAGEMAPLVL4(idx)       PAGEDIRPTR((MM_FRACTAL_BASE>>30)+((idx)&PML4_MASK))
 
-#define TMPTABLE(idx)  (*((tPAddr*)MM_TMPFRAC_BASE+((idx)&PAGE_MASK)))
-#define TMPDIR(idx)    TMPTABLE((MM_FRACTAL_BASE>>12)+((idx)&TABLE_MASK))
-#define TMPDIRPTR(idx) TMPDIR((MM_FRACTAL_BASE>>21)+((idx)&PDP_MASK))
-#define TMPMAPLVL4(idx)        TMPDIRPTR((MM_FRACTAL_BASE>>30)+((idx)&PML4_MASK))
-#define TMPCR3()       PAGETABLE(MM_TMPFRAC_BASE>>12)
+#define TMPCR3()       PAGEMAPLVL4(MM_TMPFRAC_BASE>>39)
+#define TMPTABLE(idx)  (*((Uint64*)MM_TMPFRAC_BASE+((idx)&PAGE_MASK)))
+#define TMPDIR(idx)    PAGETABLE((MM_TMPFRAC_BASE>>12)+((idx)&TABLE_MASK))
+#define TMPDIRPTR(idx) PAGEDIR((MM_TMPFRAC_BASE>>21)+((idx)&PDP_MASK))
+#define TMPMAPLVL4(idx)        PAGEDIRPTR((MM_TMPFRAC_BASE>>30)+((idx)&PML4_MASK))
 
-#define INVLPG(__addr) __asm__ __volatile__ ("invlpg (%0)"::"r"(__addr));
+#define INVLPG(__addr) __asm__ __volatile__ ("invlpg (%0)"::"r"(__addr))
+#define INVLPG_ALL()   __asm__ __volatile__ ("mov %cr3,%rax;\n\tmov %rax,%cr3;")
+#define INVLPG_GLOBAL()        __asm__ __volatile__ ("mov %cr4,%rax;\n\txorl $0x80, %eax;\n\tmov %rax,%cr4;\n\txorl $0x80, %eax;\n\tmov %rax,%cr4")
 
 // === CONSTS ===
 //tPAddr       * const gaPageTable = MM_FRACTAL_BASE;
 
-// === EXTERNS ===
+// === IMPORTS ===
+extern void    Error_Backtrace(Uint IP, Uint BP);
 extern tPAddr  gInitialPML4[512];
+extern void    Threads_SegFault(tVAddr Addr);
 
 // === PROTOTYPES ===
 void   MM_InitVirt(void);
@@ -56,6 +63,7 @@ void  MM_InitVirt(void);
 void   MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs);
 void   MM_DumpTables(tVAddr Start, tVAddr End);
  int   MM_GetPageEntryPtr(tVAddr Addr, BOOL bTemp, BOOL bAllocate, BOOL bLargePage, tPAddr **Pointer);
+ int   MM_MapEx(tVAddr VAddr, tPAddr PAddr, BOOL bTemp, BOOL bLarge);
 // int MM_Map(tVAddr VAddr, tPAddr PAddr);
 void   MM_Unmap(tVAddr VAddr);
 void   MM_ClearUser(void);
@@ -67,11 +75,12 @@ tMutex      glMM_TempFractalLock;
 // === CODE ===
 void MM_InitVirt(void)
 {
-       MM_DumpTables(0, -1L);
+//     MM_DumpTables(0, -1L);
 }
 
 void MM_FinishVirtualInit(void)
 {
+       PAGEMAPLVL4(0) = 0;
 }
 
 /**
@@ -80,24 +89,35 @@ void MM_FinishVirtualInit(void)
 void MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs)
 {
        // TODO: Implement Copy-on-Write
-       #if 0
-       if( gaPageDir  [Addr>>22] & PF_PRESENT
-        && gaPageTable[Addr>>12] & PF_PRESENT
-        && gaPageTable[Addr>>12] & PF_COW )
+       #if 1
+       if( PAGEMAPLVL4(Addr>>39) & PF_PRESENT
+        && PAGEDIRPTR (Addr>>30) & PF_PRESENT
+        && PAGEDIR    (Addr>>21) & PF_PRESENT
+        && PAGETABLE  (Addr>>12) & PF_PRESENT
+        && PAGETABLE  (Addr>>12) & PF_COW )
        {
                tPAddr  paddr;
-               if(MM_GetRefCount( gaPageTable[Addr>>12] & ~0xFFF ) == 1)
+               if(MM_GetRefCount( PAGETABLE(Addr>>12) & PADDR_MASK ) == 1)
                {
-                       gaPageTable[Addr>>12] &= ~PF_COW;
-                       gaPageTable[Addr>>12] |= PF_PRESENT|PF_WRITE;
+                       PAGETABLE(Addr>>12) &= ~PF_COW;
+                       PAGETABLE(Addr>>12) |= PF_PRESENT|PF_WRITE;
                }
                else
                {
                        //Log("MM_PageFault: COW - MM_DuplicatePage(0x%x)", Addr);
-                       paddr = MM_DuplicatePage( Addr );
-                       MM_DerefPhys( gaPageTable[Addr>>12] & ~0xFFF );
-                       gaPageTable[Addr>>12] &= PF_USER;
-                       gaPageTable[Addr>>12] |= paddr|PF_PRESENT|PF_WRITE;
+                       paddr = MM_AllocPhys();
+                       if( !paddr ) {
+                               Threads_SegFault(Addr);
+                               return ;
+                       }
+                       {
+                               void    *tmp = (void*)MM_MapTemp(paddr);
+                               memcpy( tmp, (void*)(Addr & ~0xFFF), 0x1000 );
+                               MM_FreeTemp( (tVAddr)tmp );
+                       }
+                       MM_DerefPhys( PAGETABLE(Addr>>12) & PADDR_MASK );
+                       PAGETABLE(Addr>>12) &= PF_USER;
+                       PAGETABLE(Addr>>12) |= paddr|PF_PRESENT|PF_WRITE;
                }
                
                INVLPG( Addr & ~0xFFF );
@@ -107,16 +127,15 @@ void MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs)
        
        // If it was a user, tell the thread handler
        if(ErrorCode & 4) {
-               Warning("%s %s %s memory%s",
-                       (ErrorCode&4?"User":"Kernel"),
+               Warning("User %s %s memory%s",
                        (ErrorCode&2?"write to":"read from"),
                        (ErrorCode&1?"bad/locked":"non-present"),
                        (ErrorCode&16?" (Instruction Fetch)":"")
                        );
-               Warning("User Pagefault: Instruction at %04x:%08x accessed %p",
+               Warning("User Pagefault: Instruction at %04x:%p accessed %p",
                        Regs->CS, Regs->RIP, Addr);
                __asm__ __volatile__ ("sti");   // Restart IRQs
-//             Threads_SegFault(Addr);
+               Threads_SegFault(Addr);
                return ;
        }
        
@@ -127,8 +146,7 @@ void MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs)
                Warning("Reserved Bits Trashed!");
        else
        {
-               Warning("%s %s %s memory%s",
-                       (ErrorCode&4?"User":"Kernel"),
+               Warning("Kernel %s %s memory%s",
                        (ErrorCode&2?"write to":"read from"),
                        (ErrorCode&1?"bad/locked":"non-present"),
                        (ErrorCode&16?" (Instruction Fetch)":"")
@@ -137,7 +155,7 @@ void MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs)
        
        Log("Code at %p accessed %p", Regs->RIP, Addr);
        // Print Stack Backtrace
-//     Error_Backtrace(Regs->RIP, Regs->RBP);
+       Error_Backtrace(Regs->RIP, Regs->RBP);
        
        MM_DumpTables(0, -1);
        
@@ -151,10 +169,11 @@ void MM_PageFault(tVAddr Addr, Uint ErrorCode, tRegs *Regs)
  */
 void MM_DumpTables(tVAddr Start, tVAddr End)
 {
-       const tPAddr    CHANGEABLE_BITS = 0xFF8;
+       #define CANOICAL(addr)  ((addr)&0x800000000000?(addr)|0xFFFF000000000000:(addr))
+       const tPAddr    CHANGEABLE_BITS = ~(PF_PRESENT|PF_WRITE|PF_USER|PF_COW|PF_PAGED) & 0xFFF;
        const tPAddr    MASK = ~CHANGEABLE_BITS;        // Physical address and access bits
        tVAddr  rangeStart = 0;
-       tPAddr  expected = CHANGEABLE_BITS;     // MASK is used because it's not a vaild value
+       tPAddr  expected = CHANGEABLE_BITS;     // CHANGEABLE_BITS is used because it's not a vaild value
        tVAddr  curPos;
        Uint    page;
        
@@ -177,18 +196,18 @@ void MM_DumpTables(tVAddr Start, tVAddr End)
                //Debug("&PAGETABLE(%i page) = %p", page, &PAGETABLE(page));
                
                // End of a range
-               if(
-                       !(PAGEMAPLVL4(page>>27) & PF_PRESENT)
-               ||      !(PAGEDIRPTR(page>>18) & PF_PRESENT)
-               ||      !(PAGEDIR(page>>9) & PF_PRESENT)
-               ||  !(PAGETABLE(page) & PF_PRESENT)
-               ||  (PAGETABLE(page) & MASK) != expected)
+               if(!(PAGEMAPLVL4(page>>27) & PF_PRESENT)
+               || !(PAGEDIRPTR(page>>18) & PF_PRESENT)
+               || !(PAGEDIR(page>>9) & PF_PRESENT)
+               || !(PAGETABLE(page) & PF_PRESENT)
+               || (PAGETABLE(page) & MASK) != expected)
                {                       
-                       if(expected != CHANGEABLE_BITS) {
-                               Log("%016x-0x%016x => %013x-%013x (%c%c%c%c)",
-                                       rangeStart, curPos - 1,
-                                       PAGETABLE(rangeStart>>12) & ~0xFFF,
-                                       (expected & ~0xFFF) - 1,
+                       if(expected != CHANGEABLE_BITS)
+                       {
+                               Log("%016llx => %13llx : 0x%6llx (%c%c%c%c)",
+                                       CANOICAL(rangeStart),
+                                       PAGETABLE(rangeStart>>12) & PADDR_MASK,
+                                       curPos - rangeStart,
                                        (expected & PF_PAGED ? 'p' : '-'),
                                        (expected & PF_COW ? 'C' : '-'),
                                        (expected & PF_USER ? 'U' : '-'),
@@ -224,17 +243,18 @@ void MM_DumpTables(tVAddr Start, tVAddr End)
        }
        
        if(expected != CHANGEABLE_BITS) {
-               Log("%016x-%016x => %013x-%013x (%s%s%s%s)",
-                       rangeStart, curPos - 1,
-                       PAGETABLE(rangeStart>>12) & ~0xFFF,
-                       (expected & ~0xFFF) - 1,
-                       (expected & PF_PAGED ? "p" : "-"),
-                       (expected & PF_COW ? "C" : "-"),
-                       (expected & PF_USER ? "U" : "-"),
-                       (expected & PF_WRITE ? "W" : "-")
+               Log("%016llx => %13llx : 0x%6llx (%c%c%c%c)",
+                       CANOICAL(rangeStart),
+                       PAGETABLE(rangeStart>>12) & PADDR_MASK,
+                       curPos - rangeStart,
+                       (expected & PF_PAGED ? 'p' : '-'),
+                       (expected & PF_COW ? 'C' : '-'),
+                       (expected & PF_USER ? 'U' : '-'),
+                       (expected & PF_WRITE ? 'W' : '-')
                        );
                expected = 0;
        }
+       #undef CANOICAL
 }
 
 /**
@@ -254,23 +274,29 @@ int MM_GetPageEntryPtr(tVAddr Addr, BOOL bTemp, BOOL bAllocate, BOOL bLargePage,
        const int       nADDR_SIZES = sizeof(ADDR_SIZES)/sizeof(ADDR_SIZES[0]);
         int    i;
        
+       #define BITMASK(bits)   ( (1LL << (bits))-1 )
+
        if( bTemp )
-               pmlevels[3] = (void*)MM_TMPFRAC_BASE;   // Temporary Page Table
+       {
+               pmlevels[3] = &TMPTABLE(0);     // Page Table
+               pmlevels[2] = &TMPDIR(0);       // PDIR
+               pmlevels[1] = &TMPDIRPTR(0);    // PDPT
+               pmlevels[0] = &TMPMAPLVL4(0);   // PML4
+       }
        else
+       {
                pmlevels[3] = (void*)MM_FRACTAL_BASE;   // Page Table
-       pmlevels[2] = &pmlevels[3][(MM_FRACTAL_BASE>>12)&PAGE_MASK];    // PDIR
-       pmlevels[1] = &pmlevels[2][(MM_FRACTAL_BASE>>21)&TABLE_MASK];   // PDPT
-       pmlevels[0] = &pmlevels[1][(MM_FRACTAL_BASE>>30)&PDP_MASK];     // PML4
-//     Log("pmlevels = {%p, %p, %p, %p}",
-//             MM_FRACTAL_BASE>>30, MM_FRACTAL_BASE>>21, MM_FRACTAL_BASE>>12, MM_FRACTAL_BASE);
-//     Log("pmlevels = {%p, %p, %p, %p}",
-//             pmlevels[0], pmlevels[1], pmlevels[2], pmlevels[3]);
+               pmlevels[2] = &pmlevels[3][(MM_FRACTAL_BASE>>12)&BITMASK(VIRT_BITS-12)];        // PDIR
+               pmlevels[1] = &pmlevels[2][(MM_FRACTAL_BASE>>21)&BITMASK(VIRT_BITS-21)];        // PDPT
+               pmlevels[0] = &pmlevels[1][(MM_FRACTAL_BASE>>30)&BITMASK(VIRT_BITS-30)];        // PML4
+       }
        
        // Mask address
        Addr &= (1ULL << 48)-1;
        
        for( i = 0; i < nADDR_SIZES-1; i ++ )
        {
+//             INVLPG( &pmlevels[i][ (Addr >> ADDR_SIZES[i]) & 
                
                // Check for a large page
                if( (Addr & ((1ULL << ADDR_SIZES[i])-1)) == 0 && bLargePage )
@@ -278,22 +304,25 @@ int MM_GetPageEntryPtr(tVAddr Addr, BOOL bTemp, BOOL bAllocate, BOOL bLargePage,
                        if(Pointer)     *Pointer = &pmlevels[i][Addr >> ADDR_SIZES[i]];
                        return ADDR_SIZES[i];
                }
-//             Log("&pmlevels[%i][0x%llx (>> %i)] = %p", i, Addr >> ADDR_SIZES[i], ADDR_SIZES[i],
-//                     &pmlevels[i][Addr >> ADDR_SIZES[i]]);
                // Allocate an entry if required
                if( !(pmlevels[i][Addr >> ADDR_SIZES[i]] & 1) )
                {
                        if( !bAllocate )        return -4;      // If allocation is not requested, error
-                       tmp = MM_AllocPhys();
-                       if(!tmp)        return -2;
+                       if( !(tmp = MM_AllocPhys()) )   return -2;
                        pmlevels[i][Addr >> ADDR_SIZES[i]] = tmp | 3;
-                       INVLPG( &pmlevels[i+1][ (Addr>>ADDR_SIZES[i])<<9 ] );
-                       memset( &pmlevels[i+1][ (Addr>>ADDR_SIZES[i])<<9 ], 0, 0x1000 );
+                       if( Addr < 0x800000000000 )
+                               pmlevels[i][Addr >> ADDR_SIZES[i]] |= PF_USER;
+                       INVLPG( &pmlevels[i+1][ (Addr>>ADDR_SIZES[i])*512 ] );
+                       memset( &pmlevels[i+1][ (Addr>>ADDR_SIZES[i])*512 ], 0, 0x1000 );
+                       LOG("Init PML%i ent 0x%x %p with %P", 4 - i,
+                               Addr>>ADDR_SIZES[i],
+                               (Addr>>ADDR_SIZES[i])<<ADDR_SIZES[i], tmp);
                }
+               // Catch large pages
                else if( pmlevels[i][Addr >> ADDR_SIZES[i]] & PF_LARGE )
                {
-                       if( (Addr & ((1ULL << ADDR_SIZES[i])-1)) != 0 )
-                               return -3;      // Alignment
+                       // Alignment
+                       if( (Addr & ((1ULL << ADDR_SIZES[i])-1)) != 0 ) return -3;
                        if(Pointer)     *Pointer = &pmlevels[i][Addr >> ADDR_SIZES[i]];
                        return ADDR_SIZES[i];   // Large page warning
                }
@@ -306,8 +335,12 @@ int MM_GetPageEntryPtr(tVAddr Addr, BOOL bTemp, BOOL bAllocate, BOOL bLargePage,
 
 /**
  * \brief Map a physical page to a virtual one
+ * \param VAddr        Target virtual address
+ * \param PAddr        Physical address of page
+ * \param bTemp        Use tempoary mappings
+ * \param bLarge       Treat as a large page
  */
-int MM_Map(tVAddr VAddr, tPAddr PAddr)
+int MM_MapEx(tVAddr VAddr, tPAddr PAddr, BOOL bTemp, BOOL bLarge)
 {
        tPAddr  *ent;
         int    rv;
@@ -315,20 +348,32 @@ int MM_Map(tVAddr VAddr, tPAddr PAddr)
        ENTER("xVAddr xPAddr", VAddr, PAddr);
        
        // Get page pointer (Allow allocating)
-       rv = MM_GetPageEntryPtr(VAddr, 0, 1, 0, &ent);
+       rv = MM_GetPageEntryPtr(VAddr, bTemp, 1, bLarge, &ent);
        if(rv < 0)      LEAVE_RET('i', 0);
        
-       if( *ent & 1 )
-               LEAVE_RET('i', 0);
+       if( *ent & 1 )  LEAVE_RET('i', 0);
        
        *ent = PAddr | 3;
-       
+
+       if( VAddr < 0x800000000000 )
+               *ent |= PF_USER;
+
        INVLPG( VAddr );
 
        LEAVE('i', 1);  
        return 1;
 }
 
+/**
+ * \brief Map a physical page to a virtual one
+ * \param VAddr        Target virtual address
+ * \param PAddr        Physical address of page
+ */
+int MM_Map(tVAddr VAddr, tPAddr PAddr)
+{
+       return MM_MapEx(VAddr, PAddr, 0, 0);
+}
+
 /**
  * \brief Removed a mapped page
  */
@@ -354,25 +399,13 @@ tPAddr MM_Allocate(tVAddr VAddr)
        
        ENTER("xVAddr", VAddr);
        
-       // NOTE: This is hack, but I like my dumps to be neat
-       #if 1
+       // Ensure the tables are allocated before the page (keeps things neat)
        MM_GetPageEntryPtr(VAddr, 0, 1, 0, NULL);
-       #elif 1
-       if( !MM_Map(VAddr, 0) ) // Make sure things are allocated
-       {
-               Warning("MM_Allocate: Unable to map, tables did not initialise");
-               LEAVE('i', 0);
-               return 0;
-       }
-       MM_Unmap(VAddr);
-       #endif
        
+       // Allocate the page
        ret = MM_AllocPhys();
        LOG("ret = %x", ret);
-       if(!ret) {
-               LEAVE('i', 0);
-               return 0;
-       }
+       if(!ret)        LEAVE_RET('i', 0);
        
        if( !MM_Map(VAddr, ret) )
        {
@@ -382,7 +415,7 @@ tPAddr MM_Allocate(tVAddr VAddr)
                return 0;
        }
        
-       LEAVE('x', ret);
+       LEAVE('X', ret);
        return ret;
 }
 
@@ -418,7 +451,7 @@ int MM_GetPageEntry(tVAddr Addr, tPAddr *Phys, Uint *Flags)
        ret = MM_GetPageEntryPtr(Addr, 0, 0, 0, &ptr);
        if( ret < 0 )   return 0;
        
-       *Phys = *ptr & ~0xFFF;
+       *Phys = *ptr & PADDR_MASK;
        *Flags = *ptr & 0xFFF;
        return ret;
 }
@@ -434,7 +467,9 @@ tPAddr MM_GetPhysAddr(tVAddr Addr)
        ret = MM_GetPageEntryPtr(Addr, 0, 0, 0, &ptr);
        if( ret < 0 )   return 0;
        
-       return (*ptr & ~0xFFF) | (Addr & 0xFFF);
+       if( !(*ptr & 1) )       return 0;
+       
+       return (*ptr & PADDR_MASK) | (Addr & 0xFFF);
 }
 
 /**
@@ -542,11 +577,13 @@ tVAddr MM_MapHWPages(tPAddr PAddr, Uint Number)
                }
                if( num >= 0 )  continue;
                
+               PAddr += 0x1000 * Number;
+               
                while( Number -- )
                {
                        ret -= 0x1000;
+                       PAddr -= 0x1000;
                        MM_Map(ret, PAddr);
-                       PAddr += 0x1000;
                }
                
                return ret;
@@ -620,13 +657,29 @@ tVAddr MM_AllocDMA(int Pages, int MaxBits, tPAddr *PhysAddr)
 // --- Tempory Mappings ---
 tVAddr MM_MapTemp(tPAddr PAddr)
 {
-       Log_KernelPanic("MM", "TODO: Implement MM_MapTemp");
+       const int max_slots = (MM_TMPMAP_END - MM_TMPMAP_BASE) / PAGE_SIZE;
+       tVAddr  ret = MM_TMPMAP_BASE;
+        int    i;
+       
+       for( i = 0; i < max_slots; i ++, ret += PAGE_SIZE )
+       {
+               tPAddr  *ent;
+               if( MM_GetPageEntryPtr( ret, 0, 1, 0, &ent) < 0 ) {
+                       continue ;
+               }
+
+               if( *ent & 1 )
+                       continue ;
+
+               *ent = PAddr | 3;
+               return ret;
+       }
        return 0;
 }
 
 void MM_FreeTemp(tVAddr VAddr)
 {
-       Log_KernelPanic("MM", "TODO: Implement MM_FreeTemp");
+       MM_Deallocate(VAddr);
        return ;
 }
 
@@ -635,7 +688,9 @@ void MM_FreeTemp(tVAddr VAddr)
 tPAddr MM_Clone(void)
 {
        tPAddr  ret;
-       
+        int    i;
+       tVAddr  kstackbase;
+
        // #1 Create a copy of the PML4
        ret = MM_AllocPhys();
        if(!ret)        return 0;
@@ -643,80 +698,162 @@ tPAddr MM_Clone(void)
        // #2 Alter the fractal pointer
        Mutex_Acquire(&glMM_TempFractalLock);
        TMPCR3() = ret | 3;
+       INVLPG_ALL();
        
-       INVLPG(TMPMAPLVL4(0));
-       memcpy(&TMPMAPLVL4(0), &PAGEMAPLVL4(0), 0x1000);
+       // #3 Set Copy-On-Write to all user pages
+       for( i = 0; i < 256; i ++)
+       {
+               TMPMAPLVL4(i) = PAGEMAPLVL4(i);
+//             Log_Debug("MM", "TMPMAPLVL4(%i) = 0x%016llx", i, TMPMAPLVL4(i));
+               if( TMPMAPLVL4(i) & 1 )
+               {
+                       MM_RefPhys( TMPMAPLVL4(i) & PADDR_MASK );
+                       TMPMAPLVL4(i) |= PF_COW;
+                       TMPMAPLVL4(i) &= ~PF_WRITE;
+               }
+       }
+       
+       // #4 Map in kernel pages
+       for( i = 256; i < 512; i ++ )
+       {
+               // Skip addresses:
+               // 320 0xFFFFA....      - Kernel Stacks
+               if( i == 320 )  continue;
+               // 509 0xFFFFFE0..      - Fractal mapping
+               if( i == 508 )  continue;
+               // 510 0xFFFFFE8..      - Temp fractal mapping
+               if( i == 509 )  continue;
+               
+               TMPMAPLVL4(i) = PAGEMAPLVL4(i);
+               if( TMPMAPLVL4(i) & 1 )
+                       MM_RefPhys( TMPMAPLVL4(i) & PADDR_MASK );
+       }
        
-       Log_KernelPanic("MM", "TODO: Implement MM_Clone");
+       // #5 Set fractal mapping
+       TMPMAPLVL4(508) = ret | 3;
+       TMPMAPLVL4(509) = 0;    // Temp
        
-       // #3 Set Copy-On-Write to all user pages
-       // #4 Return
+       // #6 Create kernel stack
+       //  tThread->KernelStack is the top
+       //  There is 1 guard page below the stack
+       kstackbase = Proc_GetCurThread()->KernelStack - KERNEL_STACK_SIZE;
+
+//     Log("MM_Clone: kstackbase = %p", kstackbase);
+       
+       TMPMAPLVL4(MM_KSTACK_BASE >> PML4_SHIFT) = 0;
+       for( i = 1; i < KERNEL_STACK_SIZE/0x1000; i ++ )
+       {
+               tPAddr  phys = MM_AllocPhys();
+               tVAddr  tmpmapping;
+               MM_MapEx(kstackbase+i*0x1000, phys, 1, 0);
+               
+               Log_Debug("MM", "MM_Clone: Cloning stack page %p from %P to %P",
+                       kstackbase+i*0x1000, MM_GetPhysAddr( kstackbase+i*0x1000 ), phys
+                       );
+               tmpmapping = MM_MapTemp(phys);
+               if( MM_GetPhysAddr( kstackbase+i*0x1000 ) )
+                       memcpy((void*)tmpmapping, (void*)(kstackbase+i*0x1000), 0x1000);
+               else
+                       memset((void*)tmpmapping, 0, 0x1000);
+//             if( i == 0xF )
+//                     Debug_HexDump("MM_Clone: *tmpmapping = ", (void*)tmpmapping, 0x1000);
+               MM_FreeTemp(tmpmapping);
+       }
+       
+//     MAGIC_BREAK();
+
+       // #7 Return
        TMPCR3() = 0;
-       INVLPG(TMPMAPLVL4(0));
+       INVLPG_ALL();
        Mutex_Release(&glMM_TempFractalLock);
-       return 0;
+//     Log("MM_Clone: RETURN %P", ret);
+       return ret;
 }
 
 void MM_ClearUser(void)
 {
        tVAddr  addr = 0;
-       // #1 Traverse the structure < 2^47, Deref'ing all pages
-       // #2 Free tables/dirs/pdps once they have been cleared
+        int    pml4, pdpt, pd, pt;
        
-       for( addr = 0; addr < 0x800000000000; )
+       for( pml4 = 0; pml4 < 256; pml4 ++ )
        {
-               if( PAGEMAPLVL4(addr >> PML4_SHIFT) & 1 )
+               // Catch an un-allocated PML4 entry
+               if( !(PAGEMAPLVL4(pml4) & 1) ) {
+                       addr += 1ULL << PML4_SHIFT;
+                       continue ;
+               }
+               
+               // Catch a large COW
+               if( (PAGEMAPLVL4(pml4) & PF_COW) ) {
+                       addr += 1ULL << PML4_SHIFT;
+               }
+               else
                {
-                       if( PAGEDIRPTR(addr >> PDP_SHIFT) & 1 )
+                       // TODO: Large pages
+                       
+                       // Child entries
+                       for( pdpt = 0; pdpt < 512; pdpt ++ )
                        {
-                               if( PAGEDIR(addr >> PDIR_SHIFT) & 1 )
-                               {
-                                       // Page
-                                       if( PAGETABLE(addr >> PTAB_SHIFT) & 1 ) {
-                                               MM_DerefPhys( PAGETABLE(addr >> PTAB_SHIFT) & PADDR_MASK );
-                                               PAGETABLE(addr >> PTAB_SHIFT) = 0;
-                                       }
-                                       addr += 1 << PTAB_SHIFT;
-                                       // Dereference the PDIR Entry
-                                       if( (addr + (1 << PTAB_SHIFT)) >> PDIR_SHIFT != (addr >> PDIR_SHIFT) ) {
-                                               MM_DerefPhys( PAGEMAPLVL4(addr >> PDIR_SHIFT) & PADDR_MASK );
-                                               PAGEDIR(addr >> PDIR_SHIFT) = 0;
-                                       }
-                               }
-                               else {
-                                       addr += 1 << PDIR_SHIFT;
+                               // Unallocated
+                               if( !(PAGEDIRPTR(addr >> PDP_SHIFT) & 1) ) {
+                                       addr += 1ULL << PDP_SHIFT;
                                        continue;
                                }
-                               // Dereference the PDP Entry
-                               if( (addr + (1 << PDIR_SHIFT)) >> PDP_SHIFT != (addr >> PDP_SHIFT) ) {
-                                       MM_DerefPhys( PAGEMAPLVL4(addr >> PDP_SHIFT) & PADDR_MASK );
-                                       PAGEDIRPTR(addr >> PDP_SHIFT) = 0;
+                       
+                               // Catch a large COW
+                               if( (PAGEDIRPTR(addr >> PDP_SHIFT) & PF_COW) ) {
+                                       addr += 1ULL << PDP_SHIFT;
                                }
+                               else {
+                                       // Child entries
+                                       for( pd = 0; pd < 512; pd ++ )
+                                       {
+                                               // Unallocated PDir entry
+                                               if( !(PAGEDIR(addr >> PDIR_SHIFT) & 1) ) {
+                                                       addr += 1ULL << PDIR_SHIFT;
+                                                       continue;
+                                               }
+                                               
+                                               // COW Page Table
+                                               if( PAGEDIR(addr >> PDIR_SHIFT) & PF_COW ) {
+                                                       addr += 1ULL << PDIR_SHIFT;
+                                               }
+                                               else
+                                               {
+                                                       // TODO: Catch large pages
+                                                       
+                                                       // Child entries
+                                                       for( pt = 0; pt < 512; pt ++ )
+                                                       {
+                                                               // Free page
+                                                               if( PAGETABLE(addr >> PTAB_SHIFT) & 1 ) {
+                                                                       MM_DerefPhys( PAGETABLE(addr >> PTAB_SHIFT) & PADDR_MASK );
+                                                                       PAGETABLE(addr >> PTAB_SHIFT) = 0;
+                                                               }
+                                                               addr += 1ULL << 12;
+                                                       }
+                                               }
+                                               // Free page table
+                                               MM_DerefPhys( PAGEDIR(addr >> PDIR_SHIFT) & PADDR_MASK );
+                                               PAGEDIR(addr >> PDIR_SHIFT) = 0;
+                                       }
+                               }
+                               // Free page directory
+                               MM_DerefPhys( PAGEDIRPTR(addr >> PDP_SHIFT) & PADDR_MASK );
+                               PAGEDIRPTR(addr >> PDP_SHIFT) = 0;
                        }
-                       else {
-                               addr += 1 << PDP_SHIFT;
-                               continue;
-                       }
-                       // Dereference the PML4 Entry
-                       if( (addr + (1 << PDP_SHIFT)) >> PML4_SHIFT != (addr >> PML4_SHIFT) ) {
-                               MM_DerefPhys( PAGEMAPLVL4(addr >> PML4_SHIFT) & PADDR_MASK );
-                               PAGEMAPLVL4(addr >> PML4_SHIFT) = 0;
-                       }
-               }
-               else {
-                       addr += (tVAddr)1 << PML4_SHIFT;
-                       continue;
                }
+               // Free page directory pointer table (PML4 entry)
+               MM_DerefPhys( PAGEMAPLVL4(pml4) & PADDR_MASK );
+               PAGEMAPLVL4(pml4) = 0;
        }
 }
 
-tVAddr MM_NewWorkerStack(void)
+tVAddr MM_NewWorkerStack(void *StackData, size_t StackSize)
 {
        tVAddr  ret;
         int    i;
        
-       Log_KernelPanic("MM", "TODO: Implement MM_NewWorkerStack");
-       
        // #1 Set temp fractal to PID0
        Mutex_Acquire(&glMM_TempFractalLock);
        TMPCR3() = ((tPAddr)gInitialPML4 - KERNEL_BASE) | 3;
@@ -724,7 +861,9 @@ tVAddr MM_NewWorkerStack(void)
        // #2 Scan for a free stack addresss < 2^47
        for(ret = 0x100000; ret < (1ULL << 47); ret += KERNEL_STACK_SIZE)
        {
-               if( MM_GetPhysAddr(ret) == 0 )  break;
+               tPAddr  *ptr;
+               if( MM_GetPageEntryPtr(ret, 1, 0, 0, &ptr) == 0 )       break;
+               if( !(*ptr & 1) )       break;
        }
        if( ret >= (1ULL << 47) ) {
                Mutex_Release(&glMM_TempFractalLock);
@@ -735,12 +874,31 @@ tVAddr MM_NewWorkerStack(void)
        //    - This acts as as guard page, and doesn't cost us anything.
        for( i = 0; i < KERNEL_STACK_SIZE/0x1000 - 1; i ++ )
        {
-//             MM_MapTemp
+               tPAddr  phys = MM_AllocPhys();
+               if(!phys) {
+                       // TODO: Clean up
+                       Log_Error("MM", "MM_NewWorkerStack - Unable to allocate page");
+                       return 0;
+               }
+               MM_MapEx(ret + i*0x1000, phys, 1, 0);
+       }
+
+       if( StackSize > 0x1000 ) {
+               Log_Error("MM", "MM_NewWorkerStack: StackSize(0x%x) > 0x1000, cbf handling", StackSize);
+       }
+       else {
+               tPAddr  *ptr, paddr;
+               tVAddr  tmp_addr;
+               MM_GetPageEntryPtr(ret + i*0x1000, 1, 0, 0, &ptr);
+               paddr = *ptr & ~0xFFF;
+               tmp_addr = MM_MapTemp(paddr);
+               memcpy( (void*)(tmp_addr + (0x1000 - StackSize)), StackData, StackSize );
+               MM_FreeTemp(tmp_addr);
        }
        
        Mutex_Release(&glMM_TempFractalLock);
        
-       return 0;
+       return ret + i*0x1000;
 }
 
 /**
@@ -752,11 +910,11 @@ tVAddr MM_NewKStack(void)
        Uint    i;
        for( ; base < MM_KSTACK_TOP; base += KERNEL_STACK_SIZE )
        {
-               if(MM_GetPhysAddr(base) != 0)
+               if(MM_GetPhysAddr(base+KERNEL_STACK_SIZE-0x1000) != 0)
                        continue;
                
                //Log("MM_NewKStack: Found one at %p", base + KERNEL_STACK_SIZE);
-               for( i = 0; i < KERNEL_STACK_SIZE; i += 0x1000)
+               for( i = 0x1000; i < KERNEL_STACK_SIZE; i += 0x1000)
                {
                        if( !MM_Allocate(base+i) )
                        {

UCC git Repository :: git.ucc.asn.au