These commits are when the Protocol Buffers files have changed: (only the last 100 relevant commits are shown)
| Commit: | a6001f0 | |
|---|---|---|
| Author: | Fang-Haoshu | |
add shuffle option for data_heatmap layer
The documentation is generated from this commit.
| Commit: | a731a8c | |
|---|---|---|
| Author: | Fang-Haoshu | |
merge file filler @daerduoCarey
| Commit: | dbf0b64 | |
|---|---|---|
| Author: | liuxhy237 | |
update stn, add de_transform
| Commit: | 8616f16 | |
|---|---|---|
| Author: | Fang-Haoshu | |
fix bug in prediction, add support for bbox scale
| Commit: | 6743c0c | |
|---|---|---|
| Author: | liuxhy237 | |
update files
| Commit: | eee56c8 | |
|---|---|---|
| Author: | liuxhy237 | |
add prediction heatmap layer
| Commit: | 7a9b6ab | |
|---|---|---|
| Author: | liuxhy237 | |
add detection heatmap layer
| Commit: | 3476ff4 | |
|---|---|---|
| Author: | liuxhy237 | |
final model, add UpsampleNearest filler, merge Eltwise Affine layer@ducha-aiki
| Commit: | 0d449e8 | |
|---|---|---|
| Author: | liuxhy237 | |
SHG runable
| Commit: | c8f1fbc | |
|---|---|---|
| Author: | liuxhy237 | |
add flip switch to heatmao_data_param
| Commit: | e57b2ad | |
|---|---|---|
| Author: | liuxhy237 | |
fix bug, add HG, pass simple test
| Commit: | 0963339 | |
|---|---|---|
| Author: | liuxhy237 | |
merge caffe-heatmap@tpfister
| Commit: | b351807 | |
|---|---|---|
| Author: | Fred Fang | |
| Committer: | GitHub | |
Merge pull request #1 from weiliu89/ssd Merge ssd
| Commit: | 1427713 | |
|---|---|---|
| Author: | Wei Liu | |
add demo for processing video file
| Commit: | 89eace1 | |
|---|---|---|
| Author: | Wei Liu | |
merge master and fix conflict
| Commit: | bdb9457 | |
|---|---|---|
| Author: | Alican Bozkurt | |
add default value for rms_decay
| Commit: | 5f2d845 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Add RecurrentLayer: an abstract superclass for other recurrent layer types
| Commit: | c419f85 | |
|---|---|---|
| Author: | Jonathan L Long | |
| Committer: | Jonathan L Long | |
add parameter layer for learning any bottom
| Commit: | 859cf6e | |
|---|---|---|
| Author: | Kun Wang | |
Fix an error in the example of ReshapeParameter. * this small mistake may confuse newer.
| Commit: | f87f9ac | |
|---|---|---|
| Author: | Wei Liu | |
rebase master
| Commit: | 77cde9c | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Net: setting `propagate_down: true` forces backprop
| Commit: | 251ed5a | |
|---|---|---|
| Author: | Wei Liu | |
add webcam demo
| Commit: | 3dd74cd | |
|---|---|---|
| Author: | Wei Liu | |
speed up nms and generate output for COCO
| Commit: | 519a320 | |
|---|---|---|
| Author: | Wei Liu | |
add map_object_to_agnostic to enable learning object proposal
| Commit: | 952fd17 | |
|---|---|---|
| Author: | max argus | |
| Committer: | max argus | |
Extend Crop to N-D, changed CropParameter.
| Commit: | ca9fa49 | |
|---|---|---|
| Author: | max argus | |
| Committer: | max argus | |
Crop: fixes, tests and negative axis indexing.
| Commit: | 64e78bd | |
|---|---|---|
| Author: | Jonathan L Long | |
| Committer: | max argus | |
add CropLayer: crop blob to another blob's dimensions with offsets configure offset(s) through proto definition.
| Commit: | 31a9640 | |
|---|---|---|
| Author: | Wei Liu | |
enable different choice of encoding the prior variance
| Commit: | d2ffca7 | |
|---|---|---|
| Author: | Wei Liu | |
make variance repeated in PriorBoxLayer
| Commit: | 984e302 | |
|---|---|---|
| Author: | Wei Liu | |
merge master
| Commit: | bddd04b | |
|---|---|---|
| Author: | Evan Shelhamer | |
| Committer: | Evan Shelhamer | |
deprecate input fields and upgrade automagically
| Commit: | 00598ca | |
|---|---|---|
| Author: | Evan Shelhamer | |
| Committer: | Evan Shelhamer | |
add InputLayer for Net input Create an input layer to replace oddball Net `input` fields.
| Commit: | 06cb2f4 | |
|---|---|---|
| Author: | Wei Liu | |
add more normalization to MultiBoxLoss
| Commit: | 86f9e78 | |
|---|---|---|
| Author: | Wei Liu | |
add logistic conf loss type
| Commit: | 4e2173c | |
|---|---|---|
| Author: | Wei Liu | |
add keep_top_k in DetectionOutputLayer
| Commit: | 8f847fa | |
|---|---|---|
| Author: | Youssef Kashef | |
| Committer: | Youssef Kashef | |
tranpose parameter added to IP layer to support tied weights in an autoencoder. Arguments to matrix multiplication function are conditioned on this parameter, no actual transposing takes place. test ip gradient computation with transpose on
| Commit: | a729366 | |
|---|---|---|
| Author: | Wei Liu | |
add sampler and related functions
| Commit: | ba9e3f7 | |
|---|---|---|
| Author: | Wei Liu | |
add support when evaluating on partial test data
| Commit: | a336f3e | |
|---|---|---|
| Author: | Wei Liu | |
add code_type
| Commit: | 3b7f1a7 | |
|---|---|---|
| Author: | Wei Liu | |
add variance and make max_size optional
| Commit: | c2a8dc8 | |
|---|---|---|
| Author: | Wei Liu | |
add neg_overlap for selecting hard negatives
| Commit: | 48ad4cc | |
|---|---|---|
| Author: | Wei Liu | |
do negative mining based on scores instead of overlap
| Commit: | 3342f99 | |
|---|---|---|
| Author: | Wei Liu | |
change type of size from int to float
| Commit: | 3afdc26 | |
|---|---|---|
| Author: | Wei Liu | |
add do_neg_mining in MultiBoxLossLayer
| Commit: | 32ae638 | |
|---|---|---|
| Author: | Wei Liu | |
add name_size_file in DetectionEvaluationLayer
| Commit: | 70cd366 | |
|---|---|---|
| Author: | Wei Liu | |
add SmoothL1LossLayer from Ross Girshick's Fast R-CNN
| Commit: | 5f0643d | |
|---|---|---|
| Author: | Wei Liu | |
add MaxIntegral ap_version to match VOC2012/ILSVRC AP
| Commit: | 2426d4e | |
|---|---|---|
| Author: | Wei Liu | |
add difficult property for bbox annotation
| Commit: | 89380f1 | |
|---|---|---|
| Author: | Wei Liu | |
set lr_mult to 0 instead of using fix_scale in NormalizeLayer to not learn scale parameter
| Commit: | a24f832 | |
|---|---|---|
| Author: | Wei Liu | |
add num_classes in DetectionEvaluateLayer
| Commit: | b5419e3 | |
|---|---|---|
| Author: | Wei Liu | |
add SaveOutputParameter in DetectionOutputLayer
| Commit: | 900dee1 | |
|---|---|---|
| Author: | Wei Liu | |
add NormalizeLayer from fcn branch
| Commit: | 1ae883b | |
|---|---|---|
| Author: | Wei Liu | |
add change in proto for normalize option
| Commit: | e8415b1 | |
|---|---|---|
| Author: | Wei Liu | |
add TestDetection
| Commit: | 288493d | |
|---|---|---|
| Author: | Wei Liu | |
add DetectionEvaluateLayer with test
| Commit: | 4427dac | |
|---|---|---|
| Author: | Wei Liu | |
add DetectionOutputLayer with test
| Commit: | 8c488b6 | |
|---|---|---|
| Author: | Wei Liu | |
add ApplyNMS and GetConfidenceScores to bbox_util
| Commit: | 608d0aa | |
|---|---|---|
| Author: | Wei Liu | |
fix merge upstream conflict
| Commit: | a894b40 | |
|---|---|---|
| Author: | Wei Liu | |
fix several bugs in MultiBoxLossLayer
| Commit: | b68695d | |
|---|---|---|
| Author: | Wei Liu | |
add PermuteLayer
| Commit: | 2b762b0 | |
|---|---|---|
| Author: | Wei Liu | |
fix merge upstream conflict
| Commit: | 0816907 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Separation and generalization of ChannelwiseAffineLayer into BiasLayer and ScaleLayer. The behavior of ChannelwiseAffineLayer can be reproduced by a ScaleLayer with `scale_param { bias_term: true }`. BiasLayer and ScaleLayer each take 1 or 2 bottoms, with the output having the same shape as the first. The second input -- either another bottom or a learned parameter -- will have its axes (virtually) broadcast and tiled to have the same shape as the first, after which elementwise addition (Bias) or multiplication (Scale) is performed.
| Commit: | ec04197 | |
|---|---|---|
| Author: | Dmytro Mishkin | |
| Committer: | Jeff Donahue | |
Add ChannelwiseAffine for batch norm
| Commit: | 91676b3 | |
|---|---|---|
| Author: | Wei Liu | |
put label_map_file in AnnotatedDataParameter
| Commit: | a7ac8bc | |
|---|---|---|
| Author: | Evan Shelhamer | |
Merge pull request #3388 from mohomran/exponential_linear_units Exponential Linear Units
| Commit: | bc15f86 | |
|---|---|---|
| Author: | Wei Liu | |
add MultiBoxLossLayer and bbox_util
| Commit: | 4a0c8a1 | |
|---|---|---|
| Author: | Wei Liu | |
add PriorBoxLayer which generates priors from a layer
| Commit: | 016c460 | |
|---|---|---|
| Author: | Wei Liu | |
add LabelMap and tools for create DB to store AnnotatedDatum
| Commit: | de1342f | |
|---|---|---|
| Author: | Wei Liu | |
Add AnnotatedDataLayer
| Commit: | 3e3e9ce | |
|---|---|---|
| Author: | Jonathan L Long | |
| Committer: | Jonathan L Long | |
add short description of dilation to caffe.proto
| Commit: | 93bfcb5 | |
|---|---|---|
| Author: | Fisher Yu | |
| Committer: | Jonathan L Long | |
add support for 2D dilated convolution
| Commit: | a668194 | |
|---|---|---|
| Author: | Mohamed Omran | |
| Committer: | Mohamed Omran | |
ELU layer with basic tests
| Commit: | 8b2aa70 | |
|---|---|---|
| Author: | Carl Doersch | |
| Committer: | Carl Doersch | |
Better normalization options for SoftmaxWithLoss layer.
| Commit: | 39f69fb | |
|---|---|---|
| Author: | Jeff Donahue | |
Merge pull request #3229 from cdoersch/batchnorm2 Yet another batch normalization PR
| Commit: | a52ee65 | |
|---|---|---|
| Author: | Carl Doersch | |
| Committer: | Carl Doersch | |
Cleanup batch norm layer, include global stats computation
| Commit: | 0eea815 | |
|---|---|---|
| Author: | Ronghang Hu | |
| Committer: | Ronghang Hu | |
Change solver type to string and provide solver registry
| Commit: | 321720d | |
|---|---|---|
| Author: | Evan Shelhamer | |
Merge pull request #3160 from shelhamer/cudnnV3 Basic cuDNN v3 support
| Commit: | ecac7ff | |
|---|---|---|
| Author: | Simon Layton | |
| Committer: | Evan Shelhamer | |
Initial cuDNN v3 support
| Commit: | 6c02c8b | |
|---|---|---|
| Author: | Tim Meinhardt | |
| Committer: | Tim Meinhardt | |
Add argmax_param axis
| Commit: | 9d8206e | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Im2col and Convolution layers support N spatial axes
| Commit: | 4c2ff16 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
caffe.proto: generalize ConvolutionParameter to N spatial axes
| Commit: | 251e67a | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Add TileLayer
| Commit: | 80579b8 | |
|---|---|---|
| Author: | Evan Shelhamer | |
Merge pull request #2032 from jeffdonahue/embed-layer Embed layer for lookup table of one hot encodings
| Commit: | 4e4c89b | |
|---|---|---|
| Author: | PatWie | |
| Committer: | Ronghang Hu | |
Adam solver This commit implements the Adam solver by Kingma et. al for CPU and GPU. All solver parameters are defined in the caffe.proto. This also adds an example for the MNIST dataset.
| Commit: | bb0a90e | |
|---|---|---|
| Author: | Ronghang Hu | |
Merge pull request #2903 from ronghanghu/multi_gpu Multi-GPU Data Parallelism
| Commit: | 0d34d5b | |
|---|---|---|
| Author: | Ronghang Hu | |
| Committer: | Ronghang Hu | |
Data Layers Parallel for Multi-GPU Allow data layers (and also PythonLayer when used as data layer) to be shared among worker solver's training net, and also test net for future-proof if one wants to do Multi-GPU testing. Data layers are locked during forward to ensure sequential forward.
| Commit: | 1ce3380 | |
|---|---|---|
| Author: | Mohamed Omran | |
| Committer: | Matthias Plappert | |
Implement AdaDelta; add test cases; add mnist examples
| Commit: | bcc8f50 | |
|---|---|---|
| Author: | Cyprien Noel | |
| Committer: | Evan Shelhamer | |
Add DataReader for parallel training with one DB session - Make sure each solver accesses a different subset of the data - Sequential reading of DB for performance - Prefetch a configurable amount of data to host memory - Distribute data to solvers in round-robin way for determinism
| Commit: | abe99e8 | |
|---|---|---|
| Author: | Eren Golge | |
| Committer: | Ronghang Hu | |
Implement RMSProp Solver Implement RMSProp solver and cleaned up to adjust to new solver interface that uses accumulated gradients and refactored regularization.
| Commit: | 4d299c3 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Add EmbedLayer for inner products with sparse input (one-hot vectors), with unit tests
| Commit: | 4227828 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
temporarily switch the snapshot_format default back to BINARYPROTO out of anticipation for user issues due to issue #2885, which causes Caffe to crash when it attempts to snapshot nets with duplicate layer names
| Commit: | ada055b | |
|---|---|---|
| Author: | Eric Tzeng | |
| Committer: | Eric Tzeng | |
Snapshot model weights/solver state to HDF5 files. Summary of changes: - HDF5 helper functions were moved into a separate file util/hdf5.cpp - hdf5_save_nd_dataset now saves n-d blobs, can save diffs instead of data - Minor fix for memory leak in HDF5 functions (delete instead of delete[]) - Extra methods have been added to both Net/Solver enabling snapshotting and restoring from HDF5 files - snapshot_format was added to SolverParameters, with possible values HDF5 or BINARYPROTO (default HDF5) - kMaxBlobAxes was reduced to 32 to match the limitations of HDF5
| Commit: | f973819 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Eric Tzeng | |
add double_data, double_diff to BlobProto for weights/snapshots saved when using Dtype == double
| Commit: | a756cfe | |
|---|---|---|
| Author: | Takuya Narihira | |
| Committer: | Evan Shelhamer | |
PythonLayer takes parameters by string
| Commit: | e7b2b4e | |
|---|---|---|
| Author: | philkr | |
ImageData layer default batch size of 1, and check for zero batch size
| Commit: | 823d055 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Add ReductionLayer to reduce any number of "tail" axes to a scalar value Currently implements operations SUM, MEAN, ASUM (sum of absolute values), and SUMSQ (sum of squares)
| Commit: | eb442b9 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
FlattenLayer gets a FlattenParameter with an axis, end_axis
| Commit: | 8c72fe3 | |
|---|---|---|
| Author: | Jeff Donahue | |
| Committer: | Jeff Donahue | |
Add LogLayer
| Commit: | aeef453 | |
|---|---|---|
| Author: | Evan Shelhamer | |
Merge pull request #1977 from shelhamer/accum-grad Decouple the computational batch size and minibatch size by accumulating gradients
| Commit: | 8b05a02 | |
|---|---|---|
| Author: | Jeff Donahue | |
Merge pull request #2410 from sguada/datum_transform Datum transform