Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Skeleton Of fully connected operator #2945

Merged
merged 2 commits into from
Jul 19, 2017
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
35 changes: 35 additions & 0 deletions paddle/framework/attr_checker.h
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
#include <functional>
#include <string>
#include <unordered_map>
#include <unordered_set>
#include <vector>
#include "paddle/framework/enforce.h"

Expand Down Expand Up @@ -41,6 +42,35 @@ class DefaultValueSetter {
T default_value_;
};

template <typename T>
class EnumInContainer {
public:
explicit EnumInContainer(const std::unordered_set<T>& c) : container_(c) {}
void operator()(T& val) const {
PADDLE_ENFORCE(container_.find(val) != container_.end(),
"Value %s is not in enum container %s", val,
ContainerDebugString());
}

private:
std::string ContainerDebugString() const {
std::ostringstream sout;
sout << "[";
size_t cnt = 0;
for (auto& v : container_) {
sout << v;
++cnt;
if (cnt != container_.size()) {
sout << " ,";
}
}
sout << "]";
return sout.str();
}

std::unordered_set<T> container_;
};

// check whether a certain attribute fit its limits
// an attribute can have more than one limits
template <typename T>
Expand All @@ -50,6 +80,11 @@ class TypedAttrChecker {
public:
TypedAttrChecker(const std::string& attr_name) : attr_name_(attr_name) {}

TypedAttrChecker& InEnum(const std::unordered_set<T>& range) {
value_checkers_.push_back(EnumInContainer<T>(range));
return *this;
}

TypedAttrChecker& LargerThan(const T& lower_bound) {
value_checkers_.push_back(LargerThanChecker<T>(lower_bound));
return *this;
Expand Down
6 changes: 4 additions & 2 deletions paddle/framework/net.cc
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,10 @@
namespace paddle {
namespace framework {

void PlainNet::CompleteAddOp() {
void PlainNet::CompleteAddOp(bool calc) {
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what is the meaning of cacl

add_op_done_ = true;
if (!calc) return;

std::unordered_set<std::string> input_set;
std::unordered_set<std::string> output_set;
std::unordered_set<std::string> temp_output;
Expand Down Expand Up @@ -52,7 +55,6 @@ void PlainNet::CompleteAddOp() {
}

attrs_["temporary_index"] = tmp_index;
add_op_done_ = true;
}

std::string PlainNet::DebugString() const {
Expand Down
5 changes: 2 additions & 3 deletions paddle/framework/net.h
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,6 @@ limitations under the License. */

#include <paddle/framework/op_desc.pb.h>
#include <paddle/framework/operator.h>
#include "paddle/framework/net_proto.pb.h"
#include "paddle/framework/op_proto.pb.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/scope.h"
Expand All @@ -41,7 +40,7 @@ namespace framework {
class Net : public OperatorBase {
public:
virtual void AddOp(const OperatorPtr& op) = 0;
virtual void CompleteAddOp() = 0;
virtual void CompleteAddOp(bool calc) = 0;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Declared parameter should be same with the implement.

};

using NetPtr = std::shared_ptr<Net>;
Expand Down Expand Up @@ -86,7 +85,7 @@ class PlainNet : public Net {
ops_.push_back(op);
}

void CompleteAddOp() override;
void CompleteAddOp(bool calculate = true) override;

std::string DebugString() const override;

Expand Down
6 changes: 5 additions & 1 deletion paddle/operators/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,8 @@ function(op_library TARGET)
endif()

list(LENGTH cu_srcs cu_srcs_len)
if (${cu_srcs_len} EQUAL 0)
list(LENGTH op_library_DEPS dep_len)
if (${cu_srcs_len} EQUAL 0 AND ${dep_len} EQUAL 0)
message(WARNING "The op library ${TARGET} not support GPU!")
endif()

Expand All @@ -47,3 +48,6 @@ op_library(mul_op SRCS mul_op.cc mul_op.cu)
op_library(rowwise_add_op SRCS rowwise_add_op.cu rowwise_add_op.cc)
op_library(sigmoid_op SRCS sigmoid_op.cu sigmoid_op.cc)
op_library(softmax_op SRCS softmax_op.cc softmax_op.cu)

op_library(fc_op SRCS fc_op.cc DEPS mul_op rowwise_add_op sigmoid_op
softmax_op net)
76 changes: 76 additions & 0 deletions paddle/operators/fc_op.cc
Original file line number Diff line number Diff line change
@@ -0,0 +1,76 @@
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#include "paddle/framework/net.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/operator.h"

namespace paddle {
namespace operators {

class FullyConnectedOp : public framework::PlainNet {
public:
void Init() override {
AddOp(framework::OpRegistry::CreateOp("mul",
{
Input("X"), Input("W"),
},
{Output("before_act")},
{}));
auto b = Input("b");
if (b != framework::OperatorBase::EMPTY_VAR_NAME()) {
AddOp(framework::OpRegistry::CreateOp("rowwise_add",
{Output("before_act"), Input("b")},
{Output("before_act")},
{}));
}

auto activation = GetAttr<std::string>("activation");
AddOp(framework::OpRegistry::CreateOp(
activation, {Output("before_act")}, {Output("Y")}, {}));
CompleteAddOp(false);
}
};

class FullyConnectedOpMaker : public framework::OpProtoAndCheckerMaker {
public:
FullyConnectedOpMaker(framework::OpProto *proto,
framework::OpAttrChecker *op_checker)
: OpProtoAndCheckerMaker(proto, op_checker) {
AddInput("X", "the input of fc operator");
AddInput("W", "the weight of fc operator");
AddInput("b", "the bias of fc operator");

AddOutput("Y", "the output of fc operator");
AddOutput(
"before_act", "the before activation output of fc operator", true);
AddAttr<std::string>("activation", "The activation key for fc layer")
.SetDefault("sigmoid")
.InEnum({"sigmoid", "softmax"});

//! TODO(yuyang18): Complete comment;
AddComment("FullyConnected Operator");
}
};
} // namespace operators
} // namespace paddle

USE_OP(mul);
Copy link
Member

@jacquesqiao jacquesqiao Jul 19, 2017

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if USE_OP(mul) is called in another file, the compile will fail with redefinition, so i suggest we add USE_OP in a common place.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

请教之后得知,因为 USE_OP 会在当前文件生成一个static变量,而static是private的,在不同文件中彼此不会冲突。所以不存在问题。

USE_OP(rowwise_add);
USE_OP(sigmoid);
USE_OP(softmax);

REGISTER_OP(fc,
paddle::operators::FullyConnectedOp,
paddle::operators::FullyConnectedOpMaker);
2 changes: 1 addition & 1 deletion paddle/pybind/CMakeLists.txt
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
cc_library(paddle_pybind SHARED SRCS pybind.cc DEPS pybind python
add_op mul_op rowwise_add_op sigmoid_op softmax_op)
add_op fc_op)
46 changes: 28 additions & 18 deletions paddle/pybind/pybind.cc
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@ limitations under the License. */

#include <Python.h>
#include <paddle/framework/op_registry.h>
#include <paddle/framework/operator.h>
#include <paddle/framework/scope.h>
#include <paddle/pybind/tensor_bind.h>
#include <pybind11/numpy.h>
Expand All @@ -26,10 +27,7 @@ namespace py = pybind11;
namespace pd = paddle::framework;

USE_OP(add_two);
USE_OP(softmax);
USE_OP(mul);
USE_OP(rowwise_add);
USE_OP(sigmoid);
USE_OP_WITHOUT_KERNEL(fc);

PYBIND11_PLUGIN(core) {
py::module m("core", "C++ core of Paddle Paddle");
Expand All @@ -53,7 +51,9 @@ PYBIND11_PLUGIN(core) {
self.mutable_data<int>(paddle::platform::CPUPlace());
})
.def("set", paddle::pybind::PyTensorSetFromArray<float>)
.def("set", paddle::pybind::PyTensorSetFromArray<int>);
.def("set", paddle::pybind::PyTensorSetFromArray<int>)
.def("shape",
[](pd::Tensor& self) { return pd::vectorize(self.dims()); });

py::class_<pd::Variable>(m, "Variable", R"DOC(Variable Class.

Expand Down Expand Up @@ -83,15 +83,16 @@ All parameter, weight, gradient are variables in Paddle.

//! @note: Be careful! PyBind will return std::string as an unicode, not
//! Python str. If you want a str object, you should cast them in Python.
m.def("get_all_op_protos", []() -> std::vector<std::string> {
m.def("get_all_op_protos", []() -> std::vector<py::bytes> {
auto& protos = pd::OpRegistry::protos();
std::vector<std::string> ret_values;
std::vector<py::bytes> ret_values;
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

看起来这里碰到坑了,能记录一下么?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

PyBind 11的一些使用问题

PyBind11是一个非常简单的库。这个库的使用问题记录在如下几个方面。

内存问题

Python使用GC和引用技术并存的内存问题。C++是手动管理内存。这二者之间的内存交换就有一些问题了。

对于简单类型还比较简单,直接返回值类型即可。但是对于复杂类型,就分为三种情况。

  • 由C++端负责申请,释放内存。Python只是读取这段内存。
    • 需要在注册函数的时候,添加py::return_value_policy::reference
  • 由Python端负责申请,释放内存。
    • pybind直接注册一个T* func()的函数即可,返回的对象直接使用new即可。
  • 由C++与Python共同申请,管理内存。
    • 返回std::shared_ptr即可。

字符串问题

C++中std::string不仅仅是一个字符串类型,还经常在C++端表示一段二进制内存的buffer。

而对应的Python对象有三个bytes, strunicode。所以,在使用PyBind的时候,需要根据std::string的使用场景,分别使用其他类型去接受Python参数或者返回参数给Python.

C++ Python
std::string表示二进制buffer py::bytes
std::string表示字符串 py::str 或者 std::string
std::string表示unicide字符串 std::string

C++ 端返回std::string会在C++端转换成 unicode,而C++端接受参数std::string,python端可以接受strunicode类型。

for (auto it = protos.begin(); it != protos.end(); ++it) {
PADDLE_ENFORCE(it->second.IsInitialized(),
"OpProto must all be initialized");
ret_values.emplace_back();
PADDLE_ENFORCE(it->second.SerializeToString(&ret_values.back()),
std::string str;
PADDLE_ENFORCE(it->second.SerializeToString(&str),
"Serialize OpProto Error. This could be a bug of Paddle.");
ret_values.push_back(py::bytes(str));
}
return ret_values;
});
Expand All @@ -101,17 +102,26 @@ All parameter, weight, gradient are variables in Paddle.
.def("empty", pd::OperatorBase::EMPTY_VAR_NAME)
.def("temp", pd::OperatorBase::TMP_VAR_NAME);

py::class_<paddle::platform::DeviceContext>(m, "DeviceContext")
.def_static("cpu_context", []() -> paddle::platform::DeviceContext* {
return new paddle::platform::CPUDeviceContext();
});

py::class_<pd::OperatorBase, pd::OperatorPtr>(m, "Operator")
.def("__str__", &pd::OperatorBase::DebugString)
.def_static("create", [](const std::string& protobin) {
pd::OpDesc desc;
PADDLE_ENFORCE(desc.ParsePartialFromString(protobin),
"Cannot parse user input to OpDesc");
PADDLE_ENFORCE(desc.IsInitialized(),
"User OpDesc is not initialized, reason %s",
desc.InitializationErrorString());
return pd::OpRegistry::CreateOp(desc);
});
.def_static("create",
[](py::bytes protobin) {
pd::OpDesc desc;
PADDLE_ENFORCE(desc.ParsePartialFromString(protobin),
"Cannot parse user input to OpDesc");
PADDLE_ENFORCE(desc.IsInitialized(),
"User OpDesc is not initialized, reason %s",
desc.InitializationErrorString());
return pd::OpRegistry::CreateOp(desc);
})
.def("infer_shape", &pd::OperatorBase::InferShape)
.def("run", &pd::OperatorBase::Run)
.def("outputs", [](const pd::OperatorPtr& op) { return op->outputs_; });

return m.ptr();
}
2 changes: 1 addition & 1 deletion python/paddle/v2/framework/tests/CMakeLists.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
add_python_test(test_framework test_protobuf.py test_scope.py
test_default_scope_funcs.py test_op_creation_methods.py
test_tensor.py)
test_tensor.py test_fc_op.py)
43 changes: 43 additions & 0 deletions python/paddle/v2/framework/tests/test_fc_op.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,43 @@
import paddle.v2.framework.core as core
import unittest
import numpy
import paddle.v2.framework.create_op_creation_methods as creation


class TestFc(unittest.TestCase):
def test_fc(self):
scope = core.Scope(None)
x = scope.create_var("X")
x_tensor = x.get_tensor()
x_tensor.set_dims([1000, 784])
x_tensor.alloc_float()

w = scope.create_var("W")
w_tensor = w.get_tensor()
w_tensor.set_dims([784, 100])
w_tensor.alloc_float()

w_tensor.set(numpy.random.random((784, 100)).astype("float32"))

# Set a real numpy array here.
# x_tensor.set(numpy.array([]))

op = creation.op_creations.fc(X="X", Y="Y", W="W")

for out in op.outputs():
if scope.get_var(out) is None:
scope.create_var(out).get_tensor()

tensor = scope.get_var("Y").get_tensor()
op.infer_shape(scope)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if we decide to do infer_shape in Python, or may change in the future?

self.assertEqual([1000, 100], tensor.shape())

ctx = core.DeviceContext.cpu_context()

op.run(scope, ctx)

# After complete all ops, check Y is expect or not.


if __name__ == '__main__':
unittest.main()