Skip to content

Commit

Permalink
Deploying to gh-pages from @ 9ed959f 🚀
Browse files Browse the repository at this point in the history
  • Loading branch information
facebook-github-bot committed Sep 4, 2023
1 parent 7ef363c commit 2a3be8e
Show file tree
Hide file tree
Showing 7 changed files with 45 additions and 21 deletions.
8 changes: 4 additions & 4 deletions cpp-api/split_table_batched_embeddings.html
Original file line number Diff line number Diff line change
Expand Up @@ -376,8 +376,8 @@ <h1>Table Batched Embedding Operators<a class="headerlink" href="#table-batched-
</dd></dl>

<dl class="cpp function">
<dt id="_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_t">
<span id="_CPPv342direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_t"></span><span id="_CPPv242direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_t"></span><span id="direct_mapped_lru_cache_populate_byte_cuda__at::Tensor.at::Tensor.int64_t.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.int64_t.at::Tensor.at::Tensor.int64_t"></span><span class="target" id="group__table-batched-embed-cuda_1ga18ac6da69e0b02cbfbf40f717b09f566"></span>void <code class="sig-name descname">direct_mapped_lru_cache_populate_byte_cuda</code><span class="sig-paren">(</span>at::Tensor <em>weights</em>, at::Tensor <em>hash_size_cumsum</em>, int64_t <em>total_cache_hash_size</em>, at::Tensor <em>cache_index_table_map</em>, at::Tensor <em>weights_offsets</em>, at::Tensor <em>weights_tys</em>, at::Tensor <em>D_offsets</em>, at::Tensor <em>linear_cache_indices</em>, at::Tensor <em>lxu_cache_state</em>, at::Tensor <em>lxu_cache_weights</em>, int64_t <em>time_stamp</em>, at::Tensor <em>lru_state</em>, at::Tensor <em>lxu_cache_miss_timestamp</em>, int64_t <em>row_alignment</em><span class="sig-paren">)</span><a class="headerlink" href="#_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_t" title="Permalink to this definition"></a><br /></dt>
<dt id="_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE">
<span id="_CPPv342direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE"></span><span id="_CPPv242direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE"></span><span id="direct_mapped_lru_cache_populate_byte_cuda__at::Tensor.at::Tensor.int64_t.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.at::Tensor.int64_t.at::Tensor.at::Tensor.int64_t.b.c10::optional:at::Tensor:"></span><span class="target" id="group__table-batched-embed-cuda_1gae019b6879bd9f89a146e0700d5a4bd8b"></span>void <code class="sig-name descname">direct_mapped_lru_cache_populate_byte_cuda</code><span class="sig-paren">(</span>at::Tensor <em>weights</em>, at::Tensor <em>hash_size_cumsum</em>, int64_t <em>total_cache_hash_size</em>, at::Tensor <em>cache_index_table_map</em>, at::Tensor <em>weights_offsets</em>, at::Tensor <em>weights_tys</em>, at::Tensor <em>D_offsets</em>, at::Tensor <em>linear_cache_indices</em>, at::Tensor <em>lxu_cache_state</em>, at::Tensor <em>lxu_cache_weights</em>, int64_t <em>time_stamp</em>, at::Tensor <em>lru_state</em>, at::Tensor <em>lxu_cache_miss_timestamp</em>, int64_t <em>row_alignment</em>, bool <em>gather_cache_stats</em>, c10::optional&lt;at::Tensor&gt; <em>uvm_cache_stats</em><span class="sig-paren">)</span><a class="headerlink" href="#_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE" title="Permalink to this definition"></a><br /></dt>
<dd><p>Direct-mapped (assoc=1) variant of lru_cache_populate_byte_cuda </p>
</dd></dl>

Expand All @@ -400,8 +400,8 @@ <h1>Table Batched Embedding Operators<a class="headerlink" href="#table-batched-
</dd></dl>

<dl class="cpp function">
<dt id="_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_t">
<span id="_CPPv335direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_t"></span><span id="_CPPv235direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_t"></span><span id="direct_mapped_lxu_cache_lookup_cuda__at::Tensor.at::Tensor.int64_t"></span><span class="target" id="group__table-batched-embed-cuda_1gab48d6d342bd7200b81dd3873152a0844"></span>at::Tensor <code class="sig-name descname">direct_mapped_lxu_cache_lookup_cuda</code><span class="sig-paren">(</span>at::Tensor <em>linear_cache_indices</em>, at::Tensor <em>lxu_cache_state</em>, int64_t <em>invalid_index</em><span class="sig-paren">)</span><a class="headerlink" href="#_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_t" title="Permalink to this definition"></a><br /></dt>
<dt id="_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE">
<span id="_CPPv335direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE"></span><span id="_CPPv235direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE"></span><span id="direct_mapped_lxu_cache_lookup_cuda__at::Tensor.at::Tensor.int64_t.b.c10::optional:at::Tensor:"></span><span class="target" id="group__table-batched-embed-cuda_1gab305ebdd3822794c5ac462bf5df4bb49"></span>at::Tensor <code class="sig-name descname">direct_mapped_lxu_cache_lookup_cuda</code><span class="sig-paren">(</span>at::Tensor <em>linear_cache_indices</em>, at::Tensor <em>lxu_cache_state</em>, int64_t <em>invalid_index</em>, bool <em>gather_cache_stats</em>, c10::optional&lt;at::Tensor&gt; <em>uvm_cache_stats</em><span class="sig-paren">)</span><a class="headerlink" href="#_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE" title="Permalink to this definition"></a><br /></dt>
<dd><p>Lookup the LRU/LFU cache: find the cache weights location for all indices. Look up the slots in the cache corresponding to <code class="docutils literal notranslate"><span class="pre">linear_cache_indices</span></code>, with a sentinel value for missing. </p>
</dd></dl>

Expand Down
4 changes: 2 additions & 2 deletions genindex.html
Original file line number Diff line number Diff line change
Expand Up @@ -398,9 +398,9 @@ <h2 id="D">D</h2>
</li>
</ul></td>
<td style="width: 33%; vertical-align: top;"><ul>
<li><a href="cpp-api/split_table_batched_embeddings.html#_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_t">direct_mapped_lru_cache_populate_byte_cuda (C++ function)</a>
<li><a href="cpp-api/split_table_batched_embeddings.html#_CPPv442direct_mapped_lru_cache_populate_byte_cudaN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorEN2at6TensorE7int64_tN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE">direct_mapped_lru_cache_populate_byte_cuda (C++ function)</a>
</li>
<li><a href="cpp-api/split_table_batched_embeddings.html#_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_t">direct_mapped_lxu_cache_lookup_cuda (C++ function)</a>
<li><a href="cpp-api/split_table_batched_embeddings.html#_CPPv435direct_mapped_lxu_cache_lookup_cudaN2at6TensorEN2at6TensorE7int64_tbN3c108optionalIN2at6TensorEEE">direct_mapped_lxu_cache_lookup_cuda (C++ function)</a>
</li>
</ul></td>
</tr></table>
Expand Down
44 changes: 34 additions & 10 deletions group__table-batched-embed-cuda.html
Original file line number Diff line number Diff line change
Expand Up @@ -91,16 +91,16 @@
<tr class="separator:ga00d12767ad238d73598bf7dc4d1afa06"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga5958e4cecc978d415714a3dd691fbc11"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga5958e4cecc978d415714a3dd691fbc11">lru_cache_populate_byte_cuda</a> (at::Tensor weights, at::Tensor hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor weights_tys, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, int64_t time_stamp, at::Tensor lru_state, int64_t row_alignment, bool gather_cache_stats, c10::optional&lt; at::Tensor &gt; uvm_cache_stats)</td></tr>
<tr class="separator:ga5958e4cecc978d415714a3dd691fbc11"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga18ac6da69e0b02cbfbf40f717b09f566"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga18ac6da69e0b02cbfbf40f717b09f566">direct_mapped_lru_cache_populate_byte_cuda</a> (at::Tensor weights, at::Tensor hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor weights_tys, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, int64_t time_stamp, at::Tensor lru_state, at::Tensor lxu_cache_miss_timestamp, int64_t row_alignment)</td></tr>
<tr class="separator:ga18ac6da69e0b02cbfbf40f717b09f566"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:gae019b6879bd9f89a146e0700d5a4bd8b"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#gae019b6879bd9f89a146e0700d5a4bd8b">direct_mapped_lru_cache_populate_byte_cuda</a> (at::Tensor weights, at::Tensor hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor weights_tys, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, int64_t time_stamp, at::Tensor lru_state, at::Tensor lxu_cache_miss_timestamp, int64_t row_alignment, bool gather_cache_stats, c10::optional&lt; at::Tensor &gt; uvm_cache_stats)</td></tr>
<tr class="separator:gae019b6879bd9f89a146e0700d5a4bd8b"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga854b8951ef7e78da812be97041d7d2dc"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga854b8951ef7e78da812be97041d7d2dc">lfu_cache_populate_cuda</a> (at::Tensor weights, at::Tensor cache_hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, at::Tensor lfu_state, bool stochastic_rounding)</td></tr>
<tr class="separator:ga854b8951ef7e78da812be97041d7d2dc"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga2b76a0cf452f00e77696d896d7a402f3"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga2b76a0cf452f00e77696d896d7a402f3">lfu_cache_populate_byte_cuda</a> (at::Tensor weights, at::Tensor cache_hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor weights_tys, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, at::Tensor lfu_state, int64_t row_alignment)</td></tr>
<tr class="separator:ga2b76a0cf452f00e77696d896d7a402f3"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga9ddab183e3247020b3108bfdc7d22cf9"><td class="memItemLeft" align="right" valign="top">at::Tensor&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga9ddab183e3247020b3108bfdc7d22cf9">lxu_cache_lookup_cuda</a> (at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, int64_t invalid_index, bool gather_cache_stats, c10::optional&lt; at::Tensor &gt; uvm_cache_stats)</td></tr>
<tr class="separator:ga9ddab183e3247020b3108bfdc7d22cf9"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:gab48d6d342bd7200b81dd3873152a0844"><td class="memItemLeft" align="right" valign="top">at::Tensor&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#gab48d6d342bd7200b81dd3873152a0844">direct_mapped_lxu_cache_lookup_cuda</a> (at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, int64_t invalid_index)</td></tr>
<tr class="separator:gab48d6d342bd7200b81dd3873152a0844"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:gab305ebdd3822794c5ac462bf5df4bb49"><td class="memItemLeft" align="right" valign="top">at::Tensor&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#gab305ebdd3822794c5ac462bf5df4bb49">direct_mapped_lxu_cache_lookup_cuda</a> (at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, int64_t invalid_index, bool gather_cache_stats, c10::optional&lt; at::Tensor &gt; uvm_cache_stats)</td></tr>
<tr class="separator:gab305ebdd3822794c5ac462bf5df4bb49"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:ga2b055aeb5bf2d99bfb4351271764cab1"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#ga2b055aeb5bf2d99bfb4351271764cab1">lxu_cache_flush_cuda</a> (at::Tensor uvm_weights, at::Tensor cache_hash_size_cumsum, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor D_offsets, int64_t total_D, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, bool stochastic_rounding)</td></tr>
<tr class="separator:ga2b055aeb5bf2d99bfb4351271764cab1"><td class="memSeparator" colspan="2">&#160;</td></tr>
<tr class="memitem:gaeaf8f13290f0fe389fefa3fc2a944311"><td class="memItemLeft" align="right" valign="top">void&#160;</td><td class="memItemRight" valign="bottom"><a class="el" href="group__table-batched-embed-cuda.html#gaeaf8f13290f0fe389fefa3fc2a944311">lxu_cache_locking_counter_decrement_cuda</a> (at::Tensor lxu_cache_locking_counter, at::Tensor lxu_cache_locations)</td></tr>
Expand All @@ -111,8 +111,8 @@
<a name="details" id="details"></a><h2 class="groupheader">Detailed Description</h2>
<p>The following are CUDA Operators </p>
<h2 class="groupheader">Function Documentation</h2>
<a id="ga18ac6da69e0b02cbfbf40f717b09f566" name="ga18ac6da69e0b02cbfbf40f717b09f566"></a>
<h2 class="memtitle"><span class="permalink"><a href="#ga18ac6da69e0b02cbfbf40f717b09f566">&#9670;&#160;</a></span>direct_mapped_lru_cache_populate_byte_cuda()</h2>
<a id="gae019b6879bd9f89a146e0700d5a4bd8b" name="gae019b6879bd9f89a146e0700d5a4bd8b"></a>
<h2 class="memtitle"><span class="permalink"><a href="#gae019b6879bd9f89a146e0700d5a4bd8b">&#9670;&#160;</a></span>direct_mapped_lru_cache_populate_byte_cuda()</h2>

<div class="memitem">
<div class="memproto">
Expand Down Expand Up @@ -199,7 +199,19 @@ <h2 class="memtitle"><span class="permalink"><a href="#ga18ac6da69e0b02cbfbf40f7
<td class="paramkey"></td>
<td></td>
<td class="paramtype">int64_t&#160;</td>
<td class="paramname"><em>row_alignment</em>&#160;</td>
<td class="paramname"><em>row_alignment</em>, </td>
</tr>
<tr>
<td class="paramkey"></td>
<td></td>
<td class="paramtype">bool&#160;</td>
<td class="paramname"><em>gather_cache_stats</em>, </td>
</tr>
<tr>
<td class="paramkey"></td>
<td></td>
<td class="paramtype">c10::optional&lt; at::Tensor &gt;&#160;</td>
<td class="paramname"><em>uvm_cache_stats</em>&#160;</td>
</tr>
<tr>
<td></td>
Expand All @@ -212,8 +224,8 @@ <h2 class="memtitle"><span class="permalink"><a href="#ga18ac6da69e0b02cbfbf40f7

</div>
</div>
<a id="gab48d6d342bd7200b81dd3873152a0844" name="gab48d6d342bd7200b81dd3873152a0844"></a>
<h2 class="memtitle"><span class="permalink"><a href="#gab48d6d342bd7200b81dd3873152a0844">&#9670;&#160;</a></span>direct_mapped_lxu_cache_lookup_cuda()</h2>
<a id="gab305ebdd3822794c5ac462bf5df4bb49" name="gab305ebdd3822794c5ac462bf5df4bb49"></a>
<h2 class="memtitle"><span class="permalink"><a href="#gab305ebdd3822794c5ac462bf5df4bb49">&#9670;&#160;</a></span>direct_mapped_lxu_cache_lookup_cuda()</h2>

<div class="memitem">
<div class="memproto">
Expand All @@ -234,7 +246,19 @@ <h2 class="memtitle"><span class="permalink"><a href="#gab48d6d342bd7200b81dd387
<td class="paramkey"></td>
<td></td>
<td class="paramtype">int64_t&#160;</td>
<td class="paramname"><em>invalid_index</em>&#160;</td>
<td class="paramname"><em>invalid_index</em>, </td>
</tr>
<tr>
<td class="paramkey"></td>
<td></td>
<td class="paramtype">bool&#160;</td>
<td class="paramname"><em>gather_cache_stats</em>, </td>
</tr>
<tr>
<td class="paramkey"></td>
<td></td>
<td class="paramtype">c10::optional&lt; at::Tensor &gt;&#160;</td>
<td class="paramname"><em>uvm_cache_stats</em>&#160;</td>
</tr>
<tr>
<td></td>
Expand Down
Binary file modified objects.inv
Binary file not shown.
4 changes: 2 additions & 2 deletions search/all_1.js
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
var searchData=
[
['direct_5fmapped_5flru_5fcache_5fpopulate_5fbyte_5fcuda_0',['direct_mapped_lru_cache_populate_byte_cuda',['../group__table-batched-embed-cuda.html#ga18ac6da69e0b02cbfbf40f717b09f566',1,'direct_mapped_lru_cache_populate_byte_cuda(at::Tensor weights, at::Tensor hash_size_cumsum, int64_t total_cache_hash_size, at::Tensor cache_index_table_map, at::Tensor weights_offsets, at::Tensor weights_tys, at::Tensor D_offsets, at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, at::Tensor lxu_cache_weights, int64_t time_stamp, at::Tensor lru_state, at::Tensor lxu_cache_miss_timestamp, int64_t row_alignment):&#160;split_embeddings_cache_cuda.cu'],['../group__table-batched-embed-cuda.html#ga18ac6da69e0b02cbfbf40f717b09f566',1,'direct_mapped_lru_cache_populate_byte_cuda(Tensor weights, Tensor cache_hash_size_cumsum, int64_t total_cache_hash_size, Tensor cache_index_table_map, Tensor weights_offsets, Tensor weights_tys, Tensor D_offsets, Tensor linear_cache_indices, Tensor lxu_cache_state, Tensor lxu_cache_weights, int64_t time_stamp, Tensor lru_state, Tensor lxu_cache_miss_timestamp, int64_t row_alignment):&#160;split_embeddings_cache_cuda.cu']]],
['direct_5fmapped_5flxu_5fcache_5flookup_5fcuda_1',['direct_mapped_lxu_cache_lookup_cuda',['../group__table-batched-embed-cuda.html#gab48d6d342bd7200b81dd3873152a0844',1,'direct_mapped_lxu_cache_lookup_cuda(at::Tensor linear_cache_indices, at::Tensor lxu_cache_state, int64_t invalid_index):&#160;split_embeddings_cache_cuda.cu'],['../group__table-batched-embed-cuda.html#gab48d6d342bd7200b81dd3873152a0844',1,'direct_mapped_lxu_cache_lookup_cuda(Tensor linear_cache_indices, Tensor lxu_cache_state, int64_t invalid_index):&#160;split_embeddings_cache_cuda.cu']]]
['direct_5fmapped_5flru_5fcache_5fpopulate_5fbyte_5fcuda_0',['direct_mapped_lru_cache_populate_byte_cuda',['../group__table-batched-embed-cuda.html#gae019b6879bd9f89a146e0700d5a4bd8b',1,'split_embeddings_cache_cuda.cuh']]],
['direct_5fmapped_5flxu_5fcache_5flookup_5fcuda_1',['direct_mapped_lxu_cache_lookup_cuda',['../group__table-batched-embed-cuda.html#gab305ebdd3822794c5ac462bf5df4bb49',1,'split_embeddings_cache_cuda.cuh']]]
];
Loading

0 comments on commit 2a3be8e

Please sign in to comment.