forked from apache/mxnet
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Syncing forks #2
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
…7841) * c++ dataloader and built-in image/bbox * update * fix error * fix import error * fix ci build * fix vs openmp loop type * fix warning as error with sign/unsign comp * sign/unsign comp * update to pytest * remove nose * fix tear_down * address comments * thread safe dataset * address comments * address comments * fix * serial pytest for data download
Signed-off-by: Serge Panev <spanev@nvidia.com>
* fixed overwrite of args/aux variables * fixed spacing
) * Load the user's locale before performing tests * Change the locale for the CentOS CI jobs to weed out locale-related bugs * Mark tests that fail due to the decimal point with xfail * Run localedef when generating the CentOS CI image * Cancel some Scala tests when C locale uses a non-standard decimal sep. * Rename xfail helper to xfail_when_nonstandard_decimal_separator * Fix scalastyle errors * Disable more Python tests that fail due to locale-related issues * Move assumeStandardDecimalSeparator into separate object to fix scaladoc * Disable the "symbol pow" test when running with non-standard decimal sep * Disable new tests that fail due to locale-related issues
* fix doc * fix doc * fix axis Co-authored-by: Lin <haibilin@a483e7be4c92.ant.amazon.com>
* Android build containers * ARM build containers * ARM test containers * Fix naming scheme * Set WORKDIR at correct location
* port nd.multinomial to npx.sample_categorical * move to npx.random
…at (#18263) * FFI new feature * Feature ffi x 5 * Fix pylint error * Fix pylint error * Fix around error * repeat modified
* update nvidiadocker command & remove cuda compat * replace cu101 with cuda since compat is no longer to be used * skip flaky tests * get rid of ubuntu_build_cuda and point ubuntu_cu101 to base gpu instead of cuda compat * Revert "skip flaky tests" This reverts commit 1c720fa. * revert removal of ubuntu_build_cuda * add linux gpu g4 node to all steps using g3 in unix-gpu pipeline
Co-authored-by: Lin <haibilin@a483e7be4c92.ant.amazon.com>
* Fix input gradient calculation for bidirectional LSTM For bidiractional LSTM with number of layers > 2 input gradient calculation was incorrect. Reason of wrong calculations was overwriting y derivative (dy) tensor by calculated x derivative (dx) tensor before right2left layer could use dy for own gradient calculations. Propsed fix uses additional space to avoid overwriting. * Fix gradient calculation for GRU For GRU with number of layers > 2 i2h_weight gradient for layers in the middle (all except last and first) was incorrect. Wrong caluculations were caused by assigning output pointer to input instead of calculating new input pointer. * Enable tests for GRU and LSTM gradients * Fix comments * Change loop iteration deduction * Add more test cases for fused rnn layers
* finish 5 changes * move metric.py to gluon, replace mx.metric with mx.gluon.metric in python/mxnet/ * fix importError * replace mx.metric with mx.gluon.metric in tests/python * remove global support * remove macro support * rewrite BinaryAccuracy * extend F1 to multiclass/multilabel * add tests for new F1, remove global tests * use mxnet.numpy instead of numpy * fix sanity * rewrite ce and ppl, improve some details * use mxnet.numpy.float64 * remove sklearn * remove reset_local() and get_global in other files * fix test_mlp * replace mx.metric with mx.gluon.metric in example * fix context difference * Disable -DUSE_TVM_OP on GPU builds * Fix disable tvm op for gpu runs * use label.ctx in metric.py; remove gluoncv dependency in test_cvnets * fix sanity * fix importError * remove nose Co-authored-by: Ubuntu <ubuntu@ip-172-31-12-243.us-east-2.compute.internal> Co-authored-by: Leonard Lausen <lausen@amazon.com>
TVMOP feature is now disabled on GPU builds, which caused this test to fail previously
Fix leak of ndarray objects in the frontend due to reference cycle.
…t initialized (#18306) * avoid race condition in profiler init * Update storage_profiler.h Co-authored-by: Ubuntu <ubuntu@ip-172-31-61-76.ec2.internal>
* run operator tests with naive engine * fix take tests * update skip mark * fix cuda error reset * adjust tests * disable parallel testing and naive engine for mkl/mkldnn #18244
* add dlpack functions to npx * improve tests * further improve test * fix comment
…ays built (#18308) * remove Profiler from the runtime feature list, since its always built * Update libinfo.cc * Update RunTime.pm Co-authored-by: Lin <haibilin@a483e7be4c92.ant.amazon.com>
…18588) Use CMAKE_SYSTEM_PROCESSOR to detect target architecture and make x86 related options available only when compiling for x86. Remove the code turning these options manually off on CI. Remove ANDROID cmake option which was used to decide if -lpthread needs to be specified explicitly (on most Linux systems) or not (on Android). Instead auto-detect the behavior.
* add default ctx to cachedop fwd * add test * perl fix * initial commit * update sparse tests * add aux_states * fix aux-state type * fix some tests * fix check symbolic forwrad/backward * fix symbolic grad check * arg_dict fixes * support init ops * support forward only graph * fix check symbolic backward stype * add missing file * replace extension test bind * replace bind with _bind * simplify backward_mul implementation * small fix * drop contrib.sparseembedding * remove simple_bind in test sparse ops * use simple_bind * replave simple bind in quantization * fix aux index * update amp simple_bind calls * drop ifft * fix a bug found in subgraph op * add aux_array method * replace symbols * minor fix * fix executor default context * fix import * bug fix for nd.where * add subgraph test * fix forward grad req * fix batch dot dtype * remove unused code * fix slice dtype * fix attach grad * remove tests for non-existing sparse ops * MXCachedOpGetOptimizedSymbol * fix foreach test * enhance err msg * skip failed test * add docs * add docs * fix lint * fix lint, remove quantization * fix lint * fix lint * fix lint * fix build and import * fix import * fix perl call * fix test * remove perl binding * remove reshape test * fix profiler, trt * remove tensorrt test * remove quantization tests * fix import * fix conflcit * fix lint * skip buggy test Co-authored-by: EC2 Default User <ec2-user@ip-172-31-81-80.ec2.internal> Co-authored-by: Lin <haibilin@a483e7be4c92.ant.amazon.com>
* add default ctx to cachedop fwd * add test * perl fix * initial commit * update sparse tests * add aux_states * fix aux-state type * fix some tests * fix check symbolic forwrad/backward * fix symbolic grad check * arg_dict fixes * support init ops * support forward only graph * fix check symbolic backward stype * add missing file * replace extension test bind * replace bind with _bind * simplify backward_mul implementation * small fix * drop contrib.sparseembedding * remove simple_bind in test sparse ops * use simple_bind * replave simple bind in quantization * fix aux index * update amp simple_bind calls * drop ifft * fix a bug found in subgraph op * add aux_array method * replace symbols * minor fix * fix executor default context * fix import * bug fix for nd.where * add subgraph test * fix forward grad req * fix batch dot dtype * remove unused code * fix slice dtype * fix attach grad * remove tests for non-existing sparse ops * MXCachedOpGetOptimizedSymbol * fix foreach test * enhance err msg * skip failed test * add docs * add docs * fix lint * fix lint, remove quantization * fix lint * fix lint * fix lint * fix build and import * fix import * remove scala, R, julia, perl bindings * remove cpp, matlab bindings * fix perl call * fix test * remove perl binding * remove reshape test * fix profiler, trt * remove tensorrt test * remove quantization tests * fix import * fix conflcit * fix lint * skip buggy test * remove clojure * remove executor c api * remove amalgamation * fix build * move executor folder * fix import * fix lint * fix cpp pcakge * fix predict cpp * fix cpp make * remove jnilint * remove cpp package tset * remove julia test pipeline * disable numpy tests * disable compat test for delete Co-authored-by: EC2 Default User <ec2-user@ip-172-31-81-80.ec2.internal> Co-authored-by: Lin <haibilin@a483e7be4c92.ant.amazon.com>
* redirect api reference on v-master to v1.6 * update R docs
…17949) * Initial commit of input reordering in Gluon * Add test for Gluon input reorder * Fix backward in CachedOp for input reordering * Fix test_input_reorder for backward pass * Fix merge error in NaiveCachedOp * Include correct header for std::iota Co-authored-by: Vladimir Cherepanov <vcherepanov@nvidia.com>
Update docs according to new Block APIs (#18413)
Co-authored-by: Ubuntu <ubuntu@ip-172-31-92-136.ec2.internal>
* add lans optimizer * fix * fix Co-authored-by: Zheng <shzheng@a483e789dd93.ant.amazon.com>
* adding comments explaining code optimizations * fixing broadcast_axis kernel to int32 * fixing slice_axis kernel to int32 * combining CPU and GPU implementation method signatures and cleaned up code * adding new broadcast_axis to np_matmul Co-authored-by: Rohit Kumar Srivastava <srivastava.141@buckeyemail.osu.edu>
* add numpy tril_indices ffi * Update src/api/operator/numpy/np_matrix_op.cc Co-authored-by: Haozheng Fan <hzfan9@outlook.com> Co-authored-by: Haozheng Fan <hzfan9@outlook.com>
* Add test for BatchNorm running variables synchronization * Fix BatchNorm backward synchronization It fixes issue #18610
* Fix failing empty array (log_)softmax * Modify test for npx (log_)softmax
* refactor clipboard * make lang getter more extensible * trigger ci
Add --no-pull option which disables overwriting the local docker cache based on CI docker cache. It is useful when locally changing Dockerfiles.
* flip * rollaxis * stack * fixed * retrigger ci Co-authored-by: Ubuntu <ubuntu@ip-172-31-18-97.us-east-2.compute.internal>
* fix ffi * fix less/greater error * back * submodule * fixed Co-authored-by: Ubuntu <ubuntu@ip-172-31-8-94.us-east-2.compute.internal>
* user feedback widget implementation * add user feedback widget to python docs site * update margin * add apache license * one more license * turn off feedback widget on python site * update copy * format * add event value field * turn on widget on Python site
* package created * mvn WIP * normal wip, to be tested * update * docstring added, normal mostly done * add test file * Bernoulli WIP * bernoulli wip * bernoulli doc done * dense variational WIP * add kl infra * implement normal kl method * refactor kl * add not implemented handling, rename kl_storage * add abstract method and Categorical class * rewrite logit2prob prob2logit for multiclass support * normal broadcast_to implemented * categorical mostly done * update distributions/utils.py * add dot ahead of import * fix normal F * bernoulli, normal brief tests implemented * add hybridize tests * transformation infras done * affine transformation, implemented tested * add tests cases * add sum_right_most * fix get F bug * compose transform implemented, tested * fix * add event_dim * fetch mvn from upstremm * clean code, implement normal cdf and tests * constraint in bernoulli done * fix constraint * finish half normal * add cached_property * add test on cached_property * add more features to distribution and constratins * change constraint * fix bernoulli * add independent * add independent tests * update naming of cached_property * revert * add constraints * add Cat * add Stack for imperative mode * add Stack for imperative mode * add bernoulli entropy * categorical WIP * categorical sampling implemented * finish categorical log_prob, sampling * enumerate_support finished * polish StochasticBlock, add test * add test for stochastic sequential * clean loss list in __call__ * fix affine, implement sigmoid, softmax * add gumbel, relaxed bernoulli * relaxed one-hot sampling implemented * gamma done * gamma, dirichlet implemented * beta done * gumbel softmax log-likelihood implemented * refactor tests, implement exponential, fix compose transform * weibull implemented, transformed distribution cdf icdf added * pareto implemented * uniform wip * uniform done * rewrite lgamma, implement chi2 * fix chi2 scale * F distributiion done * t implemented * fix tiny problem * cauchy done * add half cauchy * multinomial done, tests to be added * add multinomial test * MVN done, tests todo * mvn polished * fix a few precison issues * add erf, erfinv unified api and learnable transform * fix mvn attribute check * MVN done * poisson done * hack poisson for size support * geometric finished * negative binomial done * binomial done * implement some kl * add more kl * refactor kl test * add more kl * binomial kl todo * change constraint logical op implement * implement gamma entropy * finish beta dirchlet entropy * finishi all entropy * kl finished * add constraint test * domain map done * remove bayesian dense * fix tiny problems * add kl uniform normal * add kl tests * acquire patch from upstream * add some doc * finish doc * refactor kl test(WIP) * add more kl, fix float32 underflow issue * make sampling more stable * handle inconsistent mode * replace boolean idx with np.where * fix file name * add more doc * add constraint check * add half_normal/cauchy pdf cdf support check * fix import problem * change nosetest to pytest * remove buggy lines * change alias register path * attempt to fix ci * fix lint, change a few tests * fix lint * modify hybrid sequential * fix lint * change import order * add test gluon probability v2 * fix hybridize flag * change implementation of stochastic block * fix lint * fix comments * fix block * modify domain map * add raises for improper add_loss * add raises for improper add_loss * add extra cases * change collectLoss decorator to mandatory * skip stochastic block tests * remove test cases * put gpu tests back * add test_gluon_stochastic_block back * remove export test * put a test back * tiny refactor * add memory leak flag * small changes Co-authored-by: Zheng <shzheng@a483e789dd93.ant.amazon.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
(Brief description on what this PR is about)
Checklist
Essentials
Please feel free to remove inapplicable items for your PR.
Changes
Comments