Next Page: 10000
|Cache||Early diagnosis of Parkinson's disease, which causes
vital and permanent damage to both motor and non-motor symptoms, is very
important to prevent further deterioration of the patient condition. In the
present study, Parkinson's Disease data set from UCI repository is classified
using deep learning architecture. The deep learning architecture in the study
is a feed-forward neural network (FFNN) which is builded by Keras of Python.
The architecture in the study composes of an input layer, two hidden layers and
softmax function with ReLu (Rectified Linear Units) as an output layer. The
deep learning architecture solves binary classification problem since PD data
set has two classes. In order to classify the PD data set, many tests were
performed by splitting the test and train data in different ratios. The PD data
set classification was succeeded with 100% accuracy using deep learning
algorithm splitting in %20 of the data as the test and the remaining as train
data in epoch 30.|
|Cache||In the last decade, we’ve seen learning-based systems provide transformative solutions for a wide range of perception and reasoning problems, from recognizing objects in images to recognizing and translating human speech. Recent progress in deep reinforcement learning (i.e. integrating deep neural networks into reinforcement learning systems) suggests that the same kind of success could be realized in automated decision making domains. If fruitful, this line of work could allow learning-based systems to tackle active control tasks, such as robotics and autonomous driving, alongside the passive perception tasks to which they have already been successfully applied.|
|Cache||Islanding is a recent issue for power system engineers. In every normal or faulty case, the power system undergoes a sequence of states. These states are not directly observable but have a particular pattern. The patterns can be reflected in the measurements. The nonobserving states are referred to as hidden states. This paper has developed a novel hidden Markov model (HMM) based algorithm to engender a probability related to the event of islanding, depending upon the phasor measurements obtained from the smart grid. After processing the phasor data, an artificial neural network is methodically trained to provide emission probability of the hidden states. This emission probability is useful for the evaluation of the HMM. An IEEE 9 bus system is selected to test the algorithm. Several case studies are performed to generate a statistical analysis for the parameters of the HMM. The posterior probability is responsible for final pronouncement of the occurrence of islanding. The accuracy of the algorithm is estimated on the basis of finishing result. Nondetection zone is also estimated for the HMM-based islanding detection method. It is observed that this method HMM can detect islanding with less nondetection zone proficiently in a short time.|
|Cache||This paper investigates the role of modified clustering methods to account correlated uncertainties in the probabilistic power flow problem for alternate current/voltage source control-multiple terminal direct current (ac/VSC-MTdc). Such uncertainties are introduced by stochastic renewable generation units or variable loads. The wind farm is a type of stochastic generation units that grows dramatically in power networks. Due to the increase of penetration level of renewable generations, number of uncertainty sources is increased. Clustering methods are suitable tools for handling stochastic power system problems, because these methods provide acceptable results with less computation in comparison to other common methods such as Monte Carlo simulations. In this paper, two modified methods are used for clustering uncertainty sets. These methods include modified K-means algorithm, and modified self-organized map neural network. After clustering, the power flow of ac/VSC-MTdc is carried out for cluster centers of each method. Results of the probabilistic power flow for case study that is based on the IEEE 24-bus and IEEE 118-bus test systems are presented and compared with different clustering techniques, Monte Carlo simulation, and Latin hypercube sampling (LHS) methods. Such results show that power flow calculation with clustering methods is more accurate than LHS method and is faster than Monte Carlo simulation method.|
|Cache||An incredible young guest is in this episode after he attended DSGO. Edis is a 15-year-old, building his own neural networks. We discussed his background, his process of building neural networks from scratch, Kaggle competitions, and the benefit of online data science education.Additional materials: www.superdatascience.com/317|
|Cache||South African mobile data network operator, rain announced on Friday, 29 November 2019, that it has launched its intelligent 5G transport network in partnership with Huawei. Transport networks can be thought of as the “neural networks” that connect core networks to services. “With Huawei’s E2E solutions and new products, our first 5G users can experience [&hellip|
Molecules, Vol. 24, Pages 4393: Construction of Quantitative Structure Activity Relationship (QSAR) Models to Predict Potency of Structurally Diversed Janus Kinase 2 InhibitorsCache
|Janus kinase 2 (JAK2) inhibitors represent a promising therapeutic class of anticancer agents against many myeloproliferative disorders. Bioactivity data on pIC 50 of 2229 JAK2 inhibitors were employed in the construction of quantitative structure-activity relationship (QSAR) models. The models were built from 100 data splits using decision tree (DT), support vector machine (SVM), deep neural network (DNN) and random forest (RF). The predictive power of RF models were assessed via 10-fold cross validation, which afforded excellent predictive performance with R 2 and RMSE of 0.74 ± 0.05 and 0.63 ± 0.05, respectively. Moreover, test set has excellent performance of R 2 (0.75 ± 0.03) and RMSE (0.62 ± 0.04). In addition, Y-scrambling was utilized to evaluate the possibility of chance correlation of the predictive model. A thorough analysis of the substructure fingerprint count was conducted to provide insights on the inhibitory properties of JAK2 inhibitors. Molecular cluster analysis revealed that pyrazine scaffolds have nanomolar potency against JAK2.|
Sensors, Vol. 19, Pages 5277: Optimized CapsNet for Traffic Jam Speed Prediction Using Mobile Sensor Data under Urban Swarming TransportationCache
|Urban swarming transportation (UST) is a type of road transportation where multiple types of vehicles such as cars, buses, trucks, motorcycles, and bicycles, as well as pedestrians are allowed and mixed together on the roads. Predicting the traffic jam speed under UST is very different and difficult from the single road network traffic prediction which has been commonly studied in the intelligent traffic system (ITS) research. In this research, the road network wide (RNW) traffic prediction which predicts traffic jam speeds of multiple roads at once by utilizing citizens’ mobile GPS sensor records is proposed to better predict traffic jam under UST. In order to conduct the RNW traffic prediction, a specific data preprocessing is needed to convert traffic data into an image representing spatial-temporal relationships among RNW. In addition, a revised capsule network (CapsNet), named OCapsNet, which utilizes nonlinearity functions in the first two convolution layers and the modified dynamic routing to optimize the performance of CapsNet, is proposed. The experiments were conducted using real-world urban road traffic data of Jakarta to evaluate the performance. The results show that OCapsNet has better performance than Convolution Neural Network (CNN) and original CapsNet with better accuracy and precision.|
Expires November 23, 2022 23:59 PST
Buy now and get 86% off
KEY FEATURESVariational autoencoders and GANs have been two of the most interesting recent developments in deep learning and machine learning. GAN stands for generative adversarial network, where two neural networks compete with each other. Unsupervised learning means you're not trying to map input data to targets, you're just trying to learn the structure of that input data. In this course, you'll learn the structure of data in order to produce more stuff that resembles the original data.
Details & Requirements
THE EXPERTThe Lazy Programmer is a data scientist, big data engineer, and full stack software engineer. For his master's thesis he worked on brain-computer interfaces using machine learning. These assist non-verbal and non-mobile persons to communicate with their family and caregivers.
He has worked in online advertising and digital media as both a data scientist and big data engineer, and built various high-throughput web services around said data. He has created new big data pipelines using Hadoop/Pig/MapReduce, and created machine learning models to predict click-through rate, news feed recommender systems using linear regression, Bayesian Bandits, and collaborative filtering and validated the results using A/B testing.
He has taught undergraduate and graduate students in data science, statistics, machine learning, algorithms, calculus, computer graphics, and physics for students attending universities such as Columbia University, NYU, Humber College, and The New School.
Neural networks have peaked in their ability to deliver effective and meaningful results. There are four big developments that will happen with AI. A crossover with computational neuroscience will happen soon. What will be the impact of this crossover? There will be an increasing focus on how AI algorithms interact with one another.
The post Why computational neuroscience and AI will converge appeared first on JAXenter.
Leveraging Micro-Stories to Build Engagement, Inclusion, and Neural Networking in Immunology EducationCache
Storytelling is a highly effective strategy for delivering course content. It can provide real-world contexts and the relevance students desire. Through personal connections to the narrative details, anecdotes facilitate the incorporation of content into pre-existing knowledge and neural networks that enhances retention. In addition, stories can honor students' diverse backgrounds, which builds a sense of belonging and community. In turn, these aspects can drive intrinsic motivation to learn and increase students' alertness in class and overall engagement in the course. Despite the educational power of stories, there often is not enough time to integrate them into curricula. To address this dilemma, faculty can condense stories into micro-stories that require relatively minimal class time. Many aspects of stories that enhance learning can be leveraged in just a few sentences by focusing on narrative details that engage a variety of cognitive and emotional processes. In particular, the inclusion of multiple sensory descriptions and small details, like locations and names, can provide sufficient context to maintain the value stories provide. Micro-stories can function independently or extend a single theme throughout a course. Presented in this Perspective are examples of micro-stories for concepts in immunology and strategies for developing them. Proposals are made for leveraging micro-stories to enhance student engagement and course community, content retention and retrieval, and satisfaction with immunology courses of all sizes and levels.
|Cache||Neural Networks offer great promise with their ability to "create" algorithms to solve problems - without the programmer knowing how to solve the problem in the first place. Example based problem solving if you will. I would expect that if you knew precisely how to solve a particular problem to the same degree, you could certainly do it perhaps many orders of magnitude faster and possibly higher quality by coding the solution directly -- however its not always easy or practical to know such a so [...] |
Machine Learning Mortality-Classification in Clinical Documentation with Increased Accuracy in Visual-Based Analyses.Cache
Machine Learning Mortality-Classification in Clinical Documentation with Increased Accuracy in Visual-Based Analyses.
Acta Paediatr. 2019 Nov 24;:
Authors: Slattery SM, Knight DC, Weese-Mayer DE, Grobman WA, Downey DC, Murthy K
PMID: 31762098 [PubMed - as supplied by publisher]
|Cache||Open Position: An established company is looking for deep learning/AI professional for video-based action detection recognition. You will need to train Neural networks with deep learning tools and techniques to identify figures actions based on video frames sequence.|
Many scientists and engineers now use the paradigms of evolutionary computation (genetic algorithms, evolution strategies, evolutionary programming, genetic programming, classifier systems, and combinations or hybrids thereof) to tackle problems that are either intractable or unrealistically time consuming to solve through traditional computational strategies. Recently there have been vigorous initiatives to promote cross-fertilization between the EC paradigms, and also to combine these paradigms with other approaches such as neural networks to create hybrid systems with enhanced capabilities. To address the need for speedy dissemination of new ideas in these fields, and also to assist in cross-disciplinary communications and understanding, Oxford University Press and the Institute of Physics have joined forces to create a major reference publication devoted to EC fundamentals, models, algorithms and applications. This work is intended to become the standard reference resource for the evolutionary computation community. The Handbook of Evolutionary Computation will be available in loose-leaf print form, as well as in an electronic version that combines both CD-ROM and on-line (World Wide Web) access to its contents. Regularly published supplements will be available on a subscription basis. Read more...
Fresh efforts at Google to understand why an AI system says yes or no launches Explainable AI productCache
|Google has announced a new Explainable AI feature for its cloud platform, which provides more information about the features that cause an AI prediction to come up with its results. Artificial neural networks, which are used by many of today’s machine learning and AI systems, are modelled to some extent on biological brains. One of Read more about Fresh efforts at Google to understand why an AI system says yes or no launches Explainable AI product[…]|
Gabbouj, M. & Tefas, A., syyskuuta 2019, 2019 27th European Signal Processing Conference (EUSIPCO). IEEE, 5 Sivumäärä (European Signal Processing Conference).
Tutkimustuotos › › vertaisarvioitu
|Cache||In this letter, we propose a multi-level dual-attention model to classify two common macular diseases, age-related macular degeneration (AMD) and diabetic macular edema (DME) from normal macular eye conditions using optical coherence tomography (OCT) imaging technique. Our approach unifies the dual-attention mechanism at multi-levels of the pre-trained deep convolutional neural network (CNN). It provides a focused learning mechanism by taking into account both multi-level features based attention focusing on the salient coarser features and self-attention mechanism attending higher entropy regions of the finer features. Our proposed method enables the network to automatically focus on the relevant parts of the input images at different levels of feature subspaces. This leads to a more locally deformation-aware feature generation and classification. The proposed approach does not require pre-processing steps such as extraction of region of interest, denoising, and retinal flattening, making the network more robust and fully automatic. Experimental results on two macular OCT databases show the superior performance of our proposed approach as compared to the current state-of-the-art methodologies.|
|Cache||In this letter, we introduce a novel approach for nonlinear acoustic echo cancellation. The proposed approach uses the principle of transfer learning to train a neural network that approximates the nonlinear function responsible for the nonlinear distortions and generalizes this network to different acoustic conditions. The topology of the proposed network is inspired by the conventional adaptive filtering approaches for nonlinear acoustic echo cancellation. The network is trained to model the nonlinear distortions using the conventional error backpropagation algorithm. In deployment, and in order to account for any variation or discrepancy between training and deployment conditions, only a subset of the network's parameters is adapted using the significance-aware elitist resampling particle filter. The proposed approach is evaluated and verified using synthesized nonlinear distortions and real nonlinear distortions recorded by a commercial mobile phone.|
|Cache||LSTM neural networks can purportedly be used to predict crypto prices in real-time, demonstrates data scientist |
People wonder about what it would be like if a super-intelligent AI decided to place all of humanity in a realistic simulation. I wonder what it would be like if the simulation were built by today’s AI instead - whose computing power is somewhere around the level of an earthworm’s.
Specifically, what would the pies be like?
But there are now even more advanced neural nets, including GPT-2, many times larger and pretrained on millions of pages from the internet. When I used gpt-2-simple to train the smallest version of GPT-2 (145M) on the list of pies, even it was able to come up with plausible (and highly repetitive) pies, many copied from the list I gave it. I can’t blame it for being a smartass and giving me the pies right back - technically this is a perfect solution to its task of “name some pies”.
I have ways of messing with the neural net, however. One thing I can do is increase the chaos level of the text it’s generating (in this case, called the “temperature”), forcing it to go with its less-likely guesses. Another thing I can do is give it weird text to complete. When I’m asking it to generate more text, I can give it a few example pies and ask it to add to the list (the original list of pies I showed it was alphabetical, so if I give it apple pies it tends to respond with more apple pies). OR, I can give it much weirder text. If I give it a bit of Harry Potter fanfiction, it draws on its previous training and can continue the story. But, similarly to the version of GPT-2 I trained on crochet patterns, it tends to twist the stories around to its special interest. In this case, pie.
Prompt: Harry glared at Snape.
Harry glared at Snape.
Prompt: Voldemort ran his fingers gently over Darth Vader’s shoulder.
Voldemort ran his fingers gently over Darth Vader’s shoulder. “C-Custard”
Once it has steered the conversation, the neural net transitions to listing pies. Although some of them retain a certain flavor.
Voldemort Crumble Pie
And its other pies are indeed more creative, definitely not found in the original training set. They don’t sound exactly delicious to me, but maybe this advanced artificial intelligence knows something we don’t. After all, it read most of the internet, including the cooking blogs.
Hot & Sour Apple Pie
On the other hand, maybe just maybe the neural net doesn’t know what it’s doing. These seem like uniformly a bad idea.
Fog-Crust Pecan Pie
These neural net inventions miiight be okay if we knew what they actually were. At the very least, approach with caution.
And these particular examples seem to have outed themselves as the work of a less-than-whollly competent AI.
Vessel With Raisin Bran
For some examples of how the neural net managed its transitions from fanfiction to pies more or less gracefully (too long to fit in this blog post), enter your email here.
|Cache||Abstract: More than 50 years of development have made mixed integer programming (MIP) an extremely successful tool. MIP’s modeling flexibility allows it describe a wide range of business, engineering and scientific problems, and, while MIP is NP-hard, many of these problems are routinely solved in practice thanks to state-of-the-art solvers that nearly double their machine-independent speeds every year. In this talk we show how a careful application of MIP modeling techniques can lead to extremely effective MIP-based methods for three problems in machine learning and statistics.|
The first problem concerns causal inference of treatment effects in observational studies . For this problem we introduce a MIP-based matching method that directly balances covariates for multi-variate treatments and produces samples that are representative of a target population. We show how using the right MIP formulation for the problem is critical for large data sets, and illustrate the effectiveness of the resulting approach by estimating the effect that the different intensities of the 2010 Chilean earthquake had on educational outcomes. The second problem concerns the design of adaptive questionnaires for consumer preference elicitation . For this problem we introduce an approximate Bayesian method for the design of the questionnaires, which can significantly reduce the variance of the estimates obtained for certain consumer preference parameters. We show how carefully modeling the associated question selection using MIP is crucial to achieving the required near-realtime selection of the next question asked to the consumer. The third problem concerns certifying that a trained neural network is robust to adversarial attacks . For this problem we introduce strong MIP formulations that can significantly reduce the computational time needed to achieve the certification.
 Building Representative Matched Samples with Multi-valued Treatments in Large Observational Studies. M. Bennett, J. P. Vielma and J. R. Zubizarreta. Submitted for publication, 2019. arXiv:1810.06707
 Ellipsoidal methods for adaptive choice-based conjoint analysis. D. Saure and J. P. Vielma. Operations Research 67, 2019. pp. 295-597.
 Strong mixed-integer programming formulations for trained neural networks. R. Anderson, J. Huchette, C. Tjandraatmadja and J. P. Vielma. In A. Lodi and V. Nagarajan, editors, Proceedings of the 20th Conference on Integer Programming and Combinatorial Optimization (IPCO 2019), Lecture Notes in Computer Science 11480, 2019. pp. 27-42.
Bio: Juan Pablo Vielma is the Richard S. Leghorn (1939) Career Development Associate Professor at MIT Sloan School of Management and is affiliated to MIT’s Operations Research Center. Dr. Vielma has a B.S. in Mathematical Engineering from University of Chile and a Ph.D. in Industrial Engineering from the Georgia Institute of Technology. His current research interests include the theory and practice of mixed-integer mathematical optimization and applications in energy, natural resource management, marketing and statistics. In January of 2017 he was named by President Obama as one of the recipients of the Presidential Early Career Award for Scientists and Engineers (PECASE). Some of his other recognitions include the NSF CAREER Award and the INFORMS Computing Society Prize. He is currently an associate editor for Operations Research and Operations Research Letters, a member of the board of directors of the INFORMS Computing Society, and a member of the NumFocus steering committee for JuMP.
Blaize, formerly known as Thinci, unveils the first true Graph-Native silicon architecture and software platform built to process neural networks and enable AI applications with unprecedented ...
|Cache|| Hernández Munuera, Alejandro (2019). Introducing knowledge distillation in a multi-task neural network. Tesis (Master), E.T.S. de Ingenieros Informáticos (UPM) |
In this episode, I talk about different techniques that we can use to predict the outcome of some question depending on input features.
The different techniques I will go through are the ZeroR and OneR that will create a baseline for the rest of the methods.
Next up, we have the Naive Bayes classifier that is simple but powerful for some applications.
Nearest neighbor and Decision trees are next up that requires more training but is very efficient when you infer results.
Multi-layer perceptron (MLP) is the first technique that is close to the ones we usually see in Machine Learning frameworks used today. But it is just a precursor to Convolutional Neural Network (CNN) because of the size requirements. MLPs have the same size for all the hidden layers, which makes it unfeasible for larger networks.
CNNs, on the other hand, uses subsampling that will shrink the layer maps to reduce the size of the network without reducing the accuracy of the predictions.
CoinDesk • December 2, 2019, 9:00 am
Technical analysis, or the art of divining future price movements from historical data, divides opinion in the crypto world. So does does it offer real insight into the markets?
Cointelegraph.com News • December 2, 2019, 8:44 am
LSTM neural networks can purportedly be used to predict crypto prices in real-time, demonstrates data scientist
Bitcoin - The Currency of the Internet • December 2, 2019, 9:46 am
submitted by /u/cryptoadventura [link] [comments]
Bitcoin - The Currency of the Internet • December 2, 2019, 7:50 am
submitted by /u/cointastical [link] [comments]
|Cache||Haykin Neural Networks Solution|
LSTM neural networks can purportedly be used to predict crypto prices in real-time, demonstrates data scientist
Multi-language speech transcription was recently introduced into Microsoft Video Indexer at the International Broadcasters Conference (IBC). It is available as a preview capability and customers can already start experiencing it in our portal. More details on all our IBC2019 enhancements can be found here.
Multi-language videos are common media assets in the globalization context, global political summits, economic forums, and sport press conferences are examples of venues where speakers use their native language to convey their own statements. Those videos pose a unique challenge for companies that need to provide automatic transcription for video archives of large volumes. Automatic transcription technologies expect users to explicitly determine the video language in advance to convert speech to text. This manual step becomes a scalability obstacle when transcribing multi-language content as one would have to manually tag audio segments with the appropriate language.
Microsoft Video Indexer provides a unique capability of automatic spoken language identification for multi-language content. This solution allows users to easily transcribe multi-language content without going through tedious manual preparation steps before triggering it. By that, it can save anyone with large archive of videos both time and money, and enable discoverability and accessibility scenarios.
Multi-language audio transcription in Video Indexer
The multi-language transcription capability is available as part of the Video Indexer portal. Currently, it supports four languages including English, French, German and Spanish, while expecting up to three different languages in an input media asset. While uploading a new media asset you can select the “Auto-detect multi-language” option as shown below.
Additionally, each instance in the transcription section will include the language in which it was transcribed.
Customers can view the transcript and identified languages by time, jump to the specific places in the video for each language, and even see the multi-language transcription as video captions. The result transcription is also available as closed caption files (VTT, TTML, SRT, TXT, and CSV).
Language identification from an audio signal is a complex task. Acoustic environment, speaker gender, and speaker age are among a variety of factors that affect this process. We represent audio signal using a visual representation, such as spectrograms, assuming that, different languages induce unique visual patterns which can be learned using deep neural networks.
Our solution has two main stages to determine the languages used in multi-language media content. First, it employs a deep neural network to classify audio segments with very high granularity, in other words, very few seconds. While a good model will successfully identify the underlying language, it can still miss-identify some segments due to similarities between languages. Therefore, we apply a second stage for examining these misses and smooth the results accordingly.
We introduced a differentiated capability for multi-language speech transcription. With this unique capability in Video Indexer, you can become more effective about the content of your videos as it allows you to immediately start searching across videos for different language segments. During the coming few months, we will be improving this capability by adding support for more languages and improving the model’s accuracy.
For more information, visit Video Indexer’s portal or the Video Indexer developer portal, and try this new capability. Read more about the new multi-language option and how to use it in our documentation.
Ημερομηνία: Δευτέρα 25/11, Αίθουσα: Α56
Ομιλητής: Ηλίας Χαλκίδης, Υποψήφιος Διδάκτορας, Οικονομικό Πανεπιστήμιο Αθηνών
Τίτλος: Deep Neural Networks for Information Mining from Legal Texts
|Cache||Neural networks have grown from an academic curiosity to a massive industry.|
|Cache||Diffractive deep neural network is an optical machine learning framework that blends deep learning with optical diffraction and light-matter interaction to engineer diffractive surfaces that collectively perform optical computation at the speed of light. A diffractive neural network is first designed in a computer using deep learning techniques, followed by the physical fabrication of the designed layers of the neural network using e.g., 3-D printing or lithography. Since the connection between the input and output planes of a diffractive neural network is established via diffraction of light through passive layers, the inference process and the associated optical computation does not consume any power except the light used to illuminate the object of interest.|
|Cache||System combines deep neural network with programmable LED array to identify malaria parasite faster and more accurately in sparse samples. - Source: www.osa-opn.org|
Next Page: 10000
© Googlier LLC, 2019