-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RUNTIME] Simple NDArray container API in c++ #1418
Merged
Merged
Changes from all commits
Commits
Show all changes
2 commits
Select commit
Hold shift + click to select a range
File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,286 @@ | ||
/*! | ||
* Copyright (c) 2017 by Contributors | ||
* \file tvm/runtime/ndarray.h | ||
* \brief Abstract device memory management API | ||
*/ | ||
#ifndef TVM_RUNTIME_NDARRAY_H_ | ||
#define TVM_RUNTIME_NDARRAY_H_ | ||
|
||
#include <atomic> | ||
#include <vector> | ||
#include <utility> | ||
#include "./c_runtime_api.h" | ||
|
||
namespace tvm { | ||
namespace runtime { | ||
/*! | ||
* \brief Managed NDArray. | ||
* The array is backed by reference counted blocks. | ||
*/ | ||
class NDArray { | ||
public: | ||
// internal container type | ||
struct Container; | ||
/*! \brief default constructor */ | ||
NDArray() {} | ||
/*! | ||
* \brief cosntruct a NDArray that refers to data | ||
* \param data The data this NDArray refers to | ||
*/ | ||
explicit inline NDArray(Container* data); | ||
/*! | ||
* \brief copy constructor | ||
* \param other The value to be copied | ||
*/ | ||
inline NDArray(const NDArray& other); // NOLINT(*) | ||
/*! | ||
* \brief move constructor | ||
* \param other The value to be moved | ||
*/ | ||
NDArray(NDArray&& other) // NOLINT(*) | ||
: data_(other.data_) { | ||
other.data_ = nullptr; | ||
} | ||
/*! \brief destructor */ | ||
~NDArray() { | ||
this->reset(); | ||
} | ||
/*! | ||
* \brief Swap this array with another NDArray | ||
* \param other The other NDArray | ||
*/ | ||
void swap(NDArray& other) { // NOLINT(*) | ||
std::swap(data_, other.data_); | ||
} | ||
/*! | ||
* \brief copy assignmemt | ||
* \param other The value to be assigned. | ||
* \return reference to self. | ||
*/ | ||
NDArray& operator=(const NDArray& other) { // NOLINT(*) | ||
// copy-and-swap idiom | ||
NDArray(other).swap(*this); // NOLINT(*) | ||
return *this; | ||
} | ||
/*! | ||
* \brief move assignmemt | ||
* \param other The value to be assigned. | ||
* \return reference to self. | ||
*/ | ||
NDArray& operator=(NDArray&& other) { // NOLINT(*) | ||
// copy-and-swap idiom | ||
NDArray(std::move(other)).swap(*this); // NOLINT(*) | ||
return *this; | ||
} | ||
/*! \return If NDArray is defined */ | ||
bool defined() const { | ||
return data_ != nullptr; | ||
} | ||
/*! \return If both NDArray reference the same container */ | ||
bool same_as(const NDArray& other) const { | ||
return data_ == other.data_; | ||
} | ||
/*! \brief reset the content of NDArray to be nullptr */ | ||
inline void reset(); | ||
/*! | ||
* \return the reference counter | ||
* \note this number is approximate in multi-threaded setting. | ||
*/ | ||
inline int use_count() const; | ||
/*! \return Pointer to content of DLTensor */ | ||
inline const DLTensor* operator->() const; | ||
/*! | ||
* \brief Copy data content from another array. | ||
* \param other The source array to be copied from. | ||
* \note The copy may happen asynchrously if it involves a GPU context. | ||
* TVMSynchronize is necessary. | ||
*/ | ||
inline void CopyFrom(DLTensor* other); | ||
inline void CopyFrom(const NDArray& other); | ||
/*! | ||
* \brief Copy data content into another array. | ||
* \param other The source array to be copied from. | ||
* \note The copy may happen asynchrously if it involves a GPU context. | ||
* TVMSynchronize is necessary. | ||
*/ | ||
inline void CopyTo(DLTensor* other); | ||
inline void CopyTo(const NDArray& other); | ||
/*! | ||
* \brief Create a NDArray that shares the data memory with the current one. | ||
* \param shape The shape of the new array. | ||
* \param dtype The data type of the new array. | ||
* \note The memory size of new array must be smaller than the current one. | ||
*/ | ||
TVM_DLL NDArray CreateView( | ||
std::vector<int64_t> shape, DLDataType dtype); | ||
/*! | ||
* \brief Create a reference view of NDArray that | ||
* represents as DLManagedTensor. | ||
* \return A DLManagedTensor | ||
*/ | ||
TVM_DLL DLManagedTensor* ToDLPack() const; | ||
/*! | ||
* \brief Create an empty NDArray. | ||
* \param shape The shape of the new array. | ||
* \param dtype The data type of the new array. | ||
* \param ctx The context of the Array. | ||
* \return The created Array | ||
*/ | ||
TVM_DLL static NDArray Empty(std::vector<int64_t> shape, | ||
DLDataType dtype, | ||
DLContext ctx); | ||
/*! | ||
* \brief Create a NDArray backed by a dlpack tensor. | ||
* | ||
* This allows us to create a NDArray using the memory | ||
* allocated by an external deep learning framework | ||
* that is DLPack compatible. | ||
* | ||
* The memory is retained until the NDArray went out of scope. | ||
* | ||
* \return The created NDArray view. | ||
*/ | ||
TVM_DLL static NDArray FromDLPack(DLManagedTensor* tensor); | ||
/*! | ||
* \brief Function to copy data from one array to another. | ||
* \param from The source array. | ||
* \param to The target array. | ||
* \param stream The stream used in copy. | ||
*/ | ||
TVM_DLL static void CopyFromTo( | ||
DLTensor* from, DLTensor* to, TVMStreamHandle stream = nullptr); | ||
|
||
// internal namespace | ||
struct Internal; | ||
private: | ||
/*! \brief Internal Data content */ | ||
Container* data_{nullptr}; | ||
// enable internal functions | ||
friend struct Internal; | ||
friend class TVMRetValue; | ||
friend class TVMArgsSetter; | ||
}; | ||
|
||
/*! | ||
* \brief Reference counted Container object used to back NDArray. | ||
* | ||
* This object is DLTensor compatible: | ||
* the pointer to the NDArrayContainer can be directly | ||
* interpreted as a DLTensor* | ||
* | ||
* \note: do not use this function directly, use NDArray. | ||
*/ | ||
struct NDArray::Container { | ||
public: | ||
// NOTE: the first part of this structure is the same as | ||
// DLManagedTensor, note that, however, the deleter | ||
// is only called when the reference counter goes to 0 | ||
/*! | ||
* \brief The corresponding dl_tensor field. | ||
* \note it is important that the first field is DLTensor | ||
* So that this data structure is DLTensor compatible. | ||
* The head ptr of this struct can be viewed as DLTensor*. | ||
*/ | ||
DLTensor dl_tensor; | ||
/*! | ||
* \brief addtional context, reserved for recycling | ||
* \note We can attach additional content here | ||
* which the current container depend on | ||
* (e.g. reference to original memory when creating views). | ||
*/ | ||
void* manager_ctx{nullptr}; | ||
/*! | ||
* \brief Customized deleter | ||
* | ||
* \note The customized deleter is helpful to enable | ||
* different ways of memory allocator that are not | ||
* currently defined by the system. | ||
*/ | ||
void (*deleter)(Container* self) = nullptr; | ||
/*! \brief default constructor */ | ||
Container() { | ||
dl_tensor.data = nullptr; | ||
dl_tensor.ndim = 0; | ||
dl_tensor.shape = nullptr; | ||
dl_tensor.strides = nullptr; | ||
dl_tensor.byte_offset = 0; | ||
} | ||
/*! \brief developer function, increases reference counter */ | ||
void IncRef() { | ||
ref_counter_.fetch_add(1, std::memory_order_relaxed); | ||
} | ||
/*! \brief developer function, decrease reference counter */ | ||
void DecRef() { | ||
if (ref_counter_.fetch_sub(1, std::memory_order_release) == 1) { | ||
std::atomic_thread_fence(std::memory_order_acquire); | ||
if (this->deleter != nullptr) { | ||
(*this->deleter)(this); | ||
} | ||
} | ||
} | ||
|
||
private: | ||
friend class NDArray; | ||
/*! | ||
* \brief The shape container, | ||
* can be used used for shape data. | ||
*/ | ||
std::vector<int64_t> shape_; | ||
/*! \brief The internal array object */ | ||
std::atomic<int> ref_counter_{0}; | ||
}; | ||
|
||
// implementations of inline functions | ||
// the usages of functions are documented in place. | ||
inline NDArray::NDArray(Container* data) | ||
: data_(data) { | ||
data_->IncRef(); | ||
} | ||
|
||
inline NDArray::NDArray(const NDArray& other) | ||
: data_(other.data_) { | ||
data_->IncRef(); | ||
} | ||
|
||
inline void NDArray::reset() { | ||
if (data_ != nullptr) { | ||
data_->DecRef(); | ||
data_ = nullptr; | ||
} | ||
} | ||
|
||
inline void NDArray::CopyFrom(DLTensor* other) { | ||
CHECK(data_ != nullptr); | ||
CopyFromTo(other, &(data_->dl_tensor)); | ||
} | ||
|
||
inline void NDArray::CopyFrom(const NDArray& other) { | ||
CHECK(data_ != nullptr); | ||
CHECK(other.data_ != nullptr); | ||
CopyFromTo(&(other.data_->dl_tensor), &(data_->dl_tensor)); | ||
} | ||
|
||
inline void NDArray::CopyTo(DLTensor* other) { | ||
CHECK(data_ != nullptr); | ||
CopyFromTo(&(data_->dl_tensor), other); | ||
} | ||
|
||
inline void NDArray::CopyTo(const NDArray& other) { | ||
CHECK(data_ != nullptr); | ||
CHECK(other.data_ != nullptr); | ||
CopyFromTo(&(data_->dl_tensor), &(other.data_->dl_tensor)); | ||
} | ||
|
||
inline int NDArray::use_count() const { | ||
if (data_ == nullptr) return 0; | ||
return data_->ref_counter_.load(std::memory_order_relaxed); | ||
} | ||
|
||
inline const DLTensor* NDArray::operator->() const { | ||
return &(data_->dl_tensor); | ||
} | ||
|
||
} // namespace runtime | ||
} // namespace tvm | ||
|
||
#endif // TVM_RUNTIME_NDARRAY_H_ |
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we expect to share these across threads without synchronization? Atomic ref-counts are pretty expensive compared to non-atomic ones.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We might need this for multi-threaded purposes. For most cases, overhead of atomic is fine compared to overhead of other things(computing, python and kernel launch)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
👍