Faiss
 All Classes Namespaces Functions Variables Typedefs Enumerations Enumerator Friends
GpuIndexIVFFlat.cu
1 /**
2  * Copyright (c) 2015-present, Facebook, Inc.
3  * All rights reserved.
4  *
5  * This source code is licensed under the BSD+Patents license found in the
6  * LICENSE file in the root directory of this source tree.
7  */
8 
9 // Copyright 2004-present Facebook. All Rights Reserved.
10 
11 #include "GpuIndexIVFFlat.h"
12 #include "../IndexFlat.h"
13 #include "../IndexIVF.h"
14 #include "GpuIndexFlat.h"
15 #include "GpuResources.h"
16 #include "impl/IVFFlat.cuh"
17 #include "utils/CopyUtils.cuh"
18 #include "utils/DeviceUtils.h"
19 #include "utils/Float16.cuh"
20 
21 #include <limits>
22 
23 namespace faiss { namespace gpu {
24 
26  const faiss::IndexIVFFlat* index,
27  GpuIndexIVFFlatConfig config) :
28  GpuIndexIVF(resources,
29  index->d,
30  index->metric_type,
31  index->nlist,
32  config),
33  ivfFlatConfig_(config),
34  reserveMemoryVecs_(0),
35  index_(nullptr) {
36 #ifndef FAISS_USE_FLOAT16
37  FAISS_THROW_IF_NOT_MSG(!ivfFlatConfig_.useFloat16IVFStorage,
38  "float16 unsupported; need CUDA SDK >= 7.5");
39 #endif
40 
41  copyFrom(index);
42 }
43 
45  int dims,
46  int nlist,
47  faiss::MetricType metric,
48  GpuIndexIVFFlatConfig config) :
49  GpuIndexIVF(resources, dims, metric, nlist, config),
50  ivfFlatConfig_(config),
51  reserveMemoryVecs_(0),
52  index_(nullptr) {
53 
54  // faiss::Index params
55  this->is_trained = false;
56 
57 #ifndef FAISS_USE_FLOAT16
58  FAISS_THROW_IF_NOT_MSG(!ivfFlatConfig_.useFloat16IVFStorage,
59  "float16 unsupported; need CUDA SDK >= 7.5");
60 #endif
61 
62  // We haven't trained ourselves, so don't construct the IVFFlat
63  // index yet
64 }
65 
66 GpuIndexIVFFlat::~GpuIndexIVFFlat() {
67  delete index_;
68 }
69 
70 void
72  reserveMemoryVecs_ = numVecs;
73  if (index_) {
74  index_->reserveMemory(numVecs);
75  }
76 }
77 
78 void
80  DeviceScope scope(device_);
81 
82  GpuIndexIVF::copyFrom(index);
83 
84  // Clear out our old data
85  delete index_;
86  index_ = nullptr;
87 
88  // The other index might not be trained
89  if (!index->is_trained) {
90  return;
91  }
92 
93  // Otherwise, we can populate ourselves from the other index
94  this->is_trained = true;
95 
96  // Copy our lists as well
97  index_ = new IVFFlat(resources_,
99  index->metric_type == faiss::METRIC_L2,
100  ivfFlatConfig_.useFloat16IVFStorage,
101  ivfFlatConfig_.indicesOptions,
102  memorySpace_);
103 
104  FAISS_ASSERT(index->vecs.size() == index->ids.size());
105  for (size_t i = 0; i < index->vecs.size(); ++i) {
106  auto& vecs = index->vecs[i];
107  auto& ids = index->ids[i];
108 
109  FAISS_ASSERT(vecs.size() % this->d == 0);
110  auto numVecs = vecs.size() / this->d;
111  FAISS_ASSERT(numVecs == ids.size());
112 
113  // GPU index can only support max int entries per list
114  FAISS_THROW_IF_NOT_FMT(ids.size() <=
115  (size_t) std::numeric_limits<int>::max(),
116  "GPU inverted list can only support "
117  "%zu entries; %zu found",
118  (size_t) std::numeric_limits<int>::max(),
119  ids.size());
120 
121  index_->addCodeVectorsFromCpu(i, vecs.data(), ids.data(), numVecs);
122  }
123 }
124 
125 void
127  DeviceScope scope(device_);
128 
129  // We must have the indices in order to copy to ourselves
130  FAISS_THROW_IF_NOT_MSG(ivfFlatConfig_.indicesOptions != INDICES_IVF,
131  "Cannot copy to CPU as GPU index doesn't retain "
132  "indices (INDICES_IVF)");
133 
134  GpuIndexIVF::copyTo(index);
135 
136  // Clear out the old inverted lists
137  index->vecs.clear();
138  index->vecs.resize(nlist_);
139 
140  // Copy the inverted lists
141  if (index_) {
142  for (int i = 0; i < nlist_; ++i) {
143  index->vecs[i] = index_->getListVectors(i);
144  index->ids[i] = index_->getListIndices(i);
145  }
146  }
147 }
148 
149 size_t
151  if (index_) {
152  DeviceScope scope(device_);
153 
154  return index_->reclaimMemory();
155  }
156 
157  return 0;
158 }
159 
160 void
162  if (index_) {
163  DeviceScope scope(device_);
164 
165  index_->reset();
166  this->ntotal = 0;
167  } else {
168  FAISS_ASSERT(this->ntotal == 0);
169  }
170 }
171 
172 void
174  DeviceScope scope(device_);
175 
176  if (this->is_trained) {
177  FAISS_ASSERT(quantizer_->is_trained);
178  FAISS_ASSERT(quantizer_->ntotal == nlist_);
179  FAISS_ASSERT(index_);
180  return;
181  }
182 
183  FAISS_ASSERT(!index_);
184 
185  trainQuantizer_(n, x);
186 
187  // The quantizer is now trained; construct the IVF index
188  index_ = new IVFFlat(resources_,
190  this->metric_type == faiss::METRIC_L2,
191  ivfFlatConfig_.useFloat16IVFStorage,
192  ivfFlatConfig_.indicesOptions,
193  memorySpace_);
194 
195  if (reserveMemoryVecs_) {
196  index_->reserveMemory(reserveMemoryVecs_);
197  }
198 
199  this->is_trained = true;
200 }
201 
202 void
204  const float* x,
205  const Index::idx_t* xids) {
206  // Device is already set in GpuIndex::addInternal_
207  FAISS_ASSERT(index_);
208  FAISS_ASSERT(n > 0);
209 
210  auto stream = resources_->getDefaultStreamCurrentDevice();
211 
212  auto deviceVecs =
213  toDevice<float, 2>(resources_,
214  device_,
215  const_cast<float*>(x),
216  stream,
217  {(int) n, index_->getDim()});
218 
219  static_assert(sizeof(long) == sizeof(Index::idx_t), "size mismatch");
220  auto deviceIds =
221  toDevice<long, 1>(resources_,
222  device_,
223  const_cast<long*>(xids),
224  stream,
225  {(int) n});
226 
227  // Not all vectors may be able to be added (some may contain NaNs
228  // etc)
229  ntotal += index_->classifyAndAddVectors(deviceVecs, deviceIds);
230 }
231 
232 void
234  const float* x,
236  float* distances,
237  faiss::Index::idx_t* labels) const {
238  // Device is already set in GpuIndex::search
239  FAISS_ASSERT(index_);
240  FAISS_ASSERT(n > 0);
241 
242  auto stream = resources_->getDefaultStream(device_);
243 
244  // Make sure arguments are on the device we desire; use temporary
245  // memory allocations to move it if necessary
246  auto devX =
247  toDevice<float, 2>(resources_,
248  device_,
249  const_cast<float*>(x),
250  stream,
251  {(int) n, this->d});
252  auto devDistances =
253  toDevice<float, 2>(resources_,
254  device_,
255  distances,
256  stream,
257  {(int) n, (int) k});
258  auto devLabels =
259  toDevice<faiss::Index::idx_t, 2>(resources_,
260  device_,
261  labels,
262  stream,
263  {(int) n, (int) k});
264 
265  index_->query(devX, nprobe_, k, devDistances, devLabels);
266 
267  // Copy back if necessary
268  fromDevice<float, 2>(devDistances, distances, stream);
269  fromDevice<faiss::Index::idx_t, 2>(devLabels, labels, stream);
270 }
271 
272 
273 } } // namespace
GpuIndexIVFFlat(GpuResources *resources, const faiss::IndexIVFFlat *index, GpuIndexIVFFlatConfig config=GpuIndexIVFFlatConfig())
cudaStream_t getDefaultStreamCurrentDevice()
Calls getDefaultStream with the current device.
void searchImpl_(faiss::Index::idx_t n, const float *x, faiss::Index::idx_t k, float *distances, faiss::Index::idx_t *labels) const override
Called from GpuIndex for search.
int getDim() const
Return the number of dimensions we are indexing.
Definition: IVFBase.cu:100
FlatIndex * getGpuData()
For internal access.
Definition: GpuIndexFlat.h:120
void copyFrom(const faiss::IndexIVFFlat *index)
void reserveMemory(size_t numVecs)
Reserve GPU memory in our inverted lists for this number of vectors.
Definition: IVFBase.cu:45
void addImpl_(faiss::Index::idx_t n, const float *x, const faiss::Index::idx_t *ids) override
Called from GpuIndex for add/add_with_ids.
void train(Index::idx_t n, const float *x) override
std::vector< std::vector< long > > ids
Inverted lists for indexes.
Definition: IndexIVF.h:55
int d
vector dimension
Definition: Index.h:64
int classifyAndAddVectors(Tensor< float, 2, true > &vecs, Tensor< long, 1, true > &indices)
Definition: IVFFlat.cu:130
void copyTo(faiss::IndexIVFFlat *index) const
int nprobe_
Number of inverted list probes per query.
Definition: GpuIndexIVF.h:91
void reserveMemory(size_t numVecs)
Reserve GPU memory in our inverted lists for this number of vectors.
const int device_
The GPU device we are resident on.
Definition: GpuIndex.h:94
GpuResources * resources_
Manages streans, cuBLAS handles and scratch memory for devices.
Definition: GpuIndex.h:91
void copyTo(faiss::IndexIVF *index) const
Copy what we have to the CPU equivalent.
Definition: GpuIndexIVF.cu:148
long idx_t
all indices are this type
Definition: Index.h:62
int nlist_
Number of inverted lists that we manage.
Definition: GpuIndexIVF.h:88
idx_t ntotal
total nb of indexed vectors
Definition: Index.h:65
void addCodeVectorsFromCpu(int listId, const float *vecs, const long *indices, size_t numVecs)
Definition: IVFFlat.cu:58
const MemorySpace memorySpace_
The memory space of our primary storage on the GPU.
Definition: GpuIndex.h:97
GpuIndexFlat * quantizer_
Quantizer for inverted lists.
Definition: GpuIndexIVF.h:94
void query(Tensor< float, 2, true > &queries, int nprobe, int k, Tensor< float, 2, true > &outDistances, Tensor< long, 2, true > &outIndices)
Definition: IVFFlat.cu:288
std::vector< float > getListVectors(int listId) const
Return the vectors of a particular list back to the CPU.
Definition: IVFFlat.cu:354
MetricType metric_type
type of metric this index uses for search
Definition: Index.h:72
std::vector< long > getListIndices(int listId) const
Return the list indices of a particular list back to the CPU.
Definition: IVFBase.cu:207
void copyFrom(const faiss::IndexIVF *index)
Copy what we need from the CPU equivalent.
Definition: GpuIndexIVF.cu:80
bool is_trained
set if the Index does not require training, or if training is done already
Definition: Index.h:69
IndicesOptions indicesOptions
Index storage options for the GPU.
Definition: GpuIndexIVF.h:31
size_t reclaimMemory()
Definition: IVFBase.cu:105
void reset() override
removes all elements from the database.
MetricType
Some algorithms support both an inner product vetsion and a L2 search version.
Definition: Index.h:43
std::vector< std::vector< float > > vecs
Definition: IndexIVF.h:135