Cache Mapping Techniques 2 || Computer Organisation and Architecture ||Detailed Explanation
Cache Mapping Techniques 2 || Computer Organisation and Architecture ||Detailed Explanation
#cache #COA #ugc #aktu #akturesults #Computerorganisation #Cachemapping
Cache Memory is a special very high-speed memory. It is used to speed up and synchronizing with high-speed CPU. Cache memory is costlier than main memory or disk memory but economical than CPU registers. Cache memory is an extremely fast memory type that acts as a buffer between RAM and the CPU. It holds frequently requested data and instructions so that they are immediately available to the CPU when needed.
If the processor finds that the memory location is in the cache, a cache hit has occurred and data is read from cache
If the processor does not find the memory location in the cache, a cache miss has occurred. For a cache miss, the cache allocates a new entry and copies in data from main memory, then the request is fulfilled from the contents of the cache.
The performance of cache memory is frequently measured in terms of a quantity called Hit ratio.
Hit ratio = hit / (hit + miss) = no. of hits/total accesses
We can improve Cache performance using higher cache block size, higher associativity, reduce miss rate, reduce miss penalty, and reduce the time to hit in the cache.
Cache Mapping:
There are three different types of mapping used for the purpose of cache memory which are as follows: Direct mapping, Associative mapping, and Set-Associative mapping. These are explained below.
Direct Mapping –
The simplest technique, known as direct mapping, maps each block of main memory into only one possible cache line. or
In Direct mapping, assigne each memory block to a specific line in the cache. If a line is previously taken up by a memory block when a new block needs to be loaded, the old block is trashed. An address space is split into two parts index field and a tag field. The cache is used to store the tag field whereas the rest is stored in the main memory. Direct mapping`s performance is directly proportional to the Hit ratio.
i = j modulo m
where
i=cache line number
j= main memory block number
m=number of lines in the cache
For purposes of cache access, each main memory address can be viewed as consisting of three fields. The least significant w bits identify a unique word or byte within a block of main memory. In most contemporary machines, the address is at the byte level. The remaining s bits specify one of the 2s blocks of main memory. The cache logic interprets these s bits as a tag of s-r bits (most significant portion) and a line field of r bits. This latter field identifies one of the m=2r lines of the cache.
Associative Mapping –
In this type of mapping, the associative memory is used to store content and addresses of the memory word. Any block can go into any line of the cache. This means that the word id bits are used to identify which word in the block is needed, but the tag becomes all of the remaining bits. This enables the placement of any word at any place in the cache memory. It is considered to be the fastest and the most flexible mapping form.
Set-associative Mapping –
This form of mapping is an enhanced form of direct mapping where the drawbacks of direct mapping are removed. Set associative addresses the problem of possible thrashing in the direct mapping method. It does this by saying that instead of having exactly one line that a block can map to in the cache, we will group a few lines together creating a set. Then a block in memory can map to any one of the lines of a specific set..Set-associative mapping allows that each word that is present in the cache can have two or more words in the main memory for the same index address. Set associative cache mapping combines the best of direct and associative cache mapping techniques.
In this case, the cache consists of a number of sets, each of which consists of a number of lines. The relationships are
m = v * k
i= j mod v
where
i=cache set number
j=main memory block number
v=number of sets
m=number of lines in the cache number of sets
k=number of lines in each set .
Application of Cache Memory
Usually, the cache memory can store a reasonable number of blocks at any given time, but this number is small compared to the total number of blocks in the main memory.
The correspondence between the main memory blocks and those in the cache is specified by a mapping function.
Видео Cache Mapping Techniques 2 || Computer Organisation and Architecture ||Detailed Explanation канала Codemasters
#cache #COA #ugc #aktu #akturesults #Computerorganisation #Cachemapping
Cache Memory is a special very high-speed memory. It is used to speed up and synchronizing with high-speed CPU. Cache memory is costlier than main memory or disk memory but economical than CPU registers. Cache memory is an extremely fast memory type that acts as a buffer between RAM and the CPU. It holds frequently requested data and instructions so that they are immediately available to the CPU when needed.
If the processor finds that the memory location is in the cache, a cache hit has occurred and data is read from cache
If the processor does not find the memory location in the cache, a cache miss has occurred. For a cache miss, the cache allocates a new entry and copies in data from main memory, then the request is fulfilled from the contents of the cache.
The performance of cache memory is frequently measured in terms of a quantity called Hit ratio.
Hit ratio = hit / (hit + miss) = no. of hits/total accesses
We can improve Cache performance using higher cache block size, higher associativity, reduce miss rate, reduce miss penalty, and reduce the time to hit in the cache.
Cache Mapping:
There are three different types of mapping used for the purpose of cache memory which are as follows: Direct mapping, Associative mapping, and Set-Associative mapping. These are explained below.
Direct Mapping –
The simplest technique, known as direct mapping, maps each block of main memory into only one possible cache line. or
In Direct mapping, assigne each memory block to a specific line in the cache. If a line is previously taken up by a memory block when a new block needs to be loaded, the old block is trashed. An address space is split into two parts index field and a tag field. The cache is used to store the tag field whereas the rest is stored in the main memory. Direct mapping`s performance is directly proportional to the Hit ratio.
i = j modulo m
where
i=cache line number
j= main memory block number
m=number of lines in the cache
For purposes of cache access, each main memory address can be viewed as consisting of three fields. The least significant w bits identify a unique word or byte within a block of main memory. In most contemporary machines, the address is at the byte level. The remaining s bits specify one of the 2s blocks of main memory. The cache logic interprets these s bits as a tag of s-r bits (most significant portion) and a line field of r bits. This latter field identifies one of the m=2r lines of the cache.
Associative Mapping –
In this type of mapping, the associative memory is used to store content and addresses of the memory word. Any block can go into any line of the cache. This means that the word id bits are used to identify which word in the block is needed, but the tag becomes all of the remaining bits. This enables the placement of any word at any place in the cache memory. It is considered to be the fastest and the most flexible mapping form.
Set-associative Mapping –
This form of mapping is an enhanced form of direct mapping where the drawbacks of direct mapping are removed. Set associative addresses the problem of possible thrashing in the direct mapping method. It does this by saying that instead of having exactly one line that a block can map to in the cache, we will group a few lines together creating a set. Then a block in memory can map to any one of the lines of a specific set..Set-associative mapping allows that each word that is present in the cache can have two or more words in the main memory for the same index address. Set associative cache mapping combines the best of direct and associative cache mapping techniques.
In this case, the cache consists of a number of sets, each of which consists of a number of lines. The relationships are
m = v * k
i= j mod v
where
i=cache set number
j=main memory block number
v=number of sets
m=number of lines in the cache number of sets
k=number of lines in each set .
Application of Cache Memory
Usually, the cache memory can store a reasonable number of blocks at any given time, but this number is small compared to the total number of blocks in the main memory.
The correspondence between the main memory blocks and those in the cache is specified by a mapping function.
Видео Cache Mapping Techniques 2 || Computer Organisation and Architecture ||Detailed Explanation канала Codemasters
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![Cache Mapping Techniques 3 || Computer Organisation and Architecture ||Detailed Explanation](https://i.ytimg.com/vi/4jzC1owhbEw/default.jpg)
![Structures and Functions|| Data Structures and Algorithms || Saral Bhasha me](https://i.ytimg.com/vi/O5a05sUvZ6Q/default.jpg)
![Static And Dynamic Memory Allocation || Data Structures and Algorithms || Saral Bhasha me](https://i.ytimg.com/vi/wdtMD6jWisI/default.jpg)
![Time And Space Complexity || Data Structures and Algorithm || Saral Bhasha me](https://i.ytimg.com/vi/C_ZpLHnH8zc/default.jpg)
![Cache Mapping Techniques || Computer Organisation and Architecture ||Detailed Explanation](https://i.ytimg.com/vi/sPZwmzRLh7Q/default.jpg)
![Sandeep Maheshwari Inspires || Powerful Inspirational Speech](https://i.ytimg.com/vi/WjTiMawTVmM/default.jpg)
![Instruction cycle State Transition Diagram ||Computer Organisation and Architecture||Saral Bhasha me](https://i.ytimg.com/vi/miqZ5T_ialk/default.jpg)
![Floating Point Representation || Computer Architecture || Hindi](https://i.ytimg.com/vi/rAnZ0ulXytU/default.jpg)
![Arrays Basics || Data Structures and Algorithms || Saral Bhasha me](https://i.ytimg.com/vi/FeE4rsc1DS4/default.jpg)
![Shri Ram Chandra Kripalu Bhajman ll by THE Gayak Brahmin ll Shri Ram Bhumi Pujan Special](https://i.ytimg.com/vi/TTvx5DRvXWc/default.jpg)
![Hardwired Control Unit | Computer Organization and Architecture | Saral Bhasha me](https://i.ytimg.com/vi/LdqwuI4W488/default.jpg)
![Introduction to Computer Organisation and Architecture || Computer Organisation and Architecture](https://i.ytimg.com/vi/xi4vJMnzonI/default.jpg)
![Floating point representation || Computer Architecture || Hindi](https://i.ytimg.com/vi/BSHy0JUDGtU/default.jpg)
![What every Indian Student must do || Don't miss these tips by Sir Walter lewin || Physics wallah](https://i.ytimg.com/vi/FZJrdzWxSdY/default.jpg)
![Pointers to a Structure || Data Structures and Algorithms || Saral Bhasha me](https://i.ytimg.com/vi/MC3wXzpplps/default.jpg)
![Who is a Teacher || unke kartavya kya hai](https://i.ytimg.com/vi/NtyWS3x39S4/default.jpg)
![Structures in C || Data Structure and Algorithms || Saral Bhasha me](https://i.ytimg.com/vi/kMGKJeVfod4/default.jpg)
![Introduction to Data Structures || Data Structure and Algorithms || Saral Bhasha me!!!](https://i.ytimg.com/vi/_yEf2S_Tx6k/default.jpg)
![Addressing modes | Computer Organisation and architecture](https://i.ytimg.com/vi/mYm9pqum6_o/default.jpg)
![You Didn't come this far to only come this far!!!Ultimate motivation for jee,neet students!](https://i.ytimg.com/vi/hgoE8zbXCwg/default.jpg)