Skip to main content

How to construct a File System that lives in Shared Memory.



Shared Memory File System Goals

1. MOUNTED IN SHARED MEMORY

The result is a very fast, real time file system.
We use Shared Memory so that the file system is public and not private.

2. PERSISTS TO DISK

When the file system is unmounted, what happens to it?
We need to be able to save the file system so that a system reboot does not destroy it.
A great way to achieve this is to save the file system to disk.

3. EXTENSIBLE IN PLACE

We want to be able to grow the file system in place.

4. SUPPORTS CONCURRENCY

We want multiple users to be able to access the file system at the same time.
In fact, we want multiple users to be able to access the same file at the same time.
With the goals now in mind we can now talk about the major design issues:

FAT File System & Design Issues

The FAT File System has been around for quite some time. Basically it provides a pretty good file structure. But I have two problems with it:

1. FAT IS NOT EXTENSIBLE IN PLACE.

That is, you cannot shutdown the file system and then add space to the end.
You have to create a new file system and then copy in the old data.
What a pain.

2. FAT DOES NOT PROVIDE FILE LOCKING.

That is, you cannot control file concurrency access.

Preview of Part 2

Drupal ModulesThat is enough for today though. This blog post has presented the background for a Memory File System that is FAT based. Next time I’ll cover the rest of the Memory File System story. I will discuss the following subjects:
*) FAT Design.
Boot Block.
Disk Block Table.
Directory Blocks.
*) How to make the file system Extensible.
*) File Locking.

Comments

Popular posts from this blog

Python and Parquet Performance

In Pandas, PyArrow, fastparquet, AWS Data Wrangler, PySpark and Dask. This post outlines how to use all common Python libraries to read and write Parquet format while taking advantage of  columnar storage ,  columnar compression  and  data partitioning . Used together, these three optimizations can dramatically accelerate I/O for your Python applications compared to CSV, JSON, HDF or other row-based formats. Parquet makes applications possible that are simply impossible using a text format like JSON or CSV. Introduction I have recently gotten more familiar with how to work with  Parquet  datasets across the six major tools used to read and write from Parquet in the Python ecosystem:  Pandas ,  PyArrow ,  fastparquet ,  AWS Data Wrangler ,  PySpark  and  Dask . My work of late in algorithmic trading involves switching between these tools a lot and as I said I often mix up the APIs. I use Pandas and PyArrow for in-RAM comput...

Cloud computing: Update

Cloud service contracts are still too complex for many businesses to grasp the potential risks and liabilities,  Businesses are buying into cloud services without fully understanding what they're paying for and what they can expect from the service. "One of the big barriers to using cloud computing is a lack of trust. I think you should be able to know what you're getting and what it means — and it should be easy to ensure that the terms in your contract are reasonable: open, transparent, safe and fair. Even if you don’t have a law degree," "Sensible, plain language contracts" be designed to spell out clear service level agreements and what a businesses' rights are on a range of issues, such as which third parties would be able to access a businesses' information or whether a firm will be notified in the event of data being stolen. Drawing up model contracts for cloud services is a "key pillar" of the  Cloud Computing strategy ....