From 240a49e68a6da5c79f558245e4a9f45fad66fc01 Mon Sep 17 00:00:00 2001 From: Muhammed Fatih BALIN Date: Tue, 27 Aug 2024 14:15:32 -0400 Subject: [PATCH] [GraphBolt][Doc] Fix typo in documentation. (#7751) --- python/dgl/graphbolt/impl/cpu_cached_feature.py | 2 +- python/dgl/graphbolt/impl/gpu_cached_feature.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/python/dgl/graphbolt/impl/cpu_cached_feature.py b/python/dgl/graphbolt/impl/cpu_cached_feature.py index 184c382fc3c7..0cb437494b13 100644 --- a/python/dgl/graphbolt/impl/cpu_cached_feature.py +++ b/python/dgl/graphbolt/impl/cpu_cached_feature.py @@ -17,7 +17,7 @@ class CPUCachedFeature(Feature): - r"""CPU cached feature wrapping a fallback feature. Use `cpu_feature_cache` + r"""CPU cached feature wrapping a fallback feature. Use `cpu_cached_feature` to construct an instance of this class. Parameters diff --git a/python/dgl/graphbolt/impl/gpu_cached_feature.py b/python/dgl/graphbolt/impl/gpu_cached_feature.py index 7c7b01cb38ef..59a66413843c 100644 --- a/python/dgl/graphbolt/impl/gpu_cached_feature.py +++ b/python/dgl/graphbolt/impl/gpu_cached_feature.py @@ -18,7 +18,7 @@ class GPUCachedFeature(Feature): r"""GPU cached feature wrapping a fallback feature. It uses the least recently used (LRU) algorithm as the cache eviction policy. Use - `gpu_feature_cache` to construct an instance of this class. + `gpu_cached_feature` to construct an instance of this class. Places the GPU cache to torch.cuda.current_device().