Juyang (John) Weng | |
Nationality: | Chinese-American |
Occupation: | Computer engineer, neuroscientist, author, and academic |
Education: | BSc., Computer Science MSc., Computer Science PhD., Computer Science |
Alma Mater: | Fudan University University of Illinois at Urbana-Champaign |
Thesis Year: | 1989 |
Doctoral Advisor: | Thomas S. Huang Narendra Ahuja |
Workplaces: | Brain-Mind Institute GENISAMA Michigan State University |
Juyang (John) Weng is a Chinese-American computer engineer, neuroscientist, author, and academic. He is a former professor at the Department of Computer Science and Engineering at Michigan State University and the President of Brain-Mind Institute and GENISAMA.[1] Weng has conducted research on grounded machine learning by intersecting computer science and engineering, with brain and cognitive science. In collaborative research efforts with his coworkers, he has explored mental architectures and computational models for autonomous development across various domains such as vision, audition, touch, behaviors, and motivational systems, both in biological and engineered systems. He has authored two books, Natural and Artificial Intelligence: Introduction to Computational Brain-Mind and Motion and Structure from Image Sequences, and is the editor of the book series 'New Frontiers in Robotics.' In addition, he has published over 300 articles.
Weng is a Life Fellow of the Institute of Electrical and Electronics Engineers (IEEE) and the Founder and President of the Brain-Mind Institute, and the startup GENISAMA. He is also the Founder and Editor-in-chief of the International Journal of Humanoid Robotics and the Brain-Mind Magazine and an Associate Editor of the IEEE Transactions on Autonomous Mental Development (now Cognitive and Developmental Systems).[2] Additionally, he served as a Guest editor for five special issues, including What AI and Neuroscience Can Learn from Each Other: Open Problems in Models and Theories, Cognitive Computation,[3] The Special Issue on Brain Imaging-informed Multimodal Analysis, IEEE Transactions on Autonomous Mental Development,[4] and The Special Issue on Autonomous Mental Development, International Journal of Humanoid Robotics.[5]
Weng obtained his BS degree from Fudan University in 1982, followed by earning his M.Sc. and Ph.D. degrees in computer science from the University of Illinois at Urbana-Champaign in 1985 and 1989, respectively.[6]
Following his Ph.D., in 1990, Weng began his academic career as a visiting assistant research professor at the Beckman Institute of the University of Illinois, Urbana. From 1992 to 1998, he served as an assistant professor at Michigan State University, becoming associate professor in 1998 and professor in 2003. Retired now[7]
Weng's research revolves around grounded machine learning, spanning vision, audition, natural language understanding, planning, and real-time hardware implementations. He is also involved in technology transfer through his startup, GENISAMA, which focuses on grounded, emergent, natural, incremental, skull-closed, attentive, motivated, and abstract systems. His theoretical contributions include mathematically proving that Developmental Networks (DNs) he developed can learn any universal turing machines and establishing a theory on Autonomous Programming For General Purposes (APFGP), supporting Conscious Machine Learning.[8] [9] Weng has worked on developmental networks from Cresceptron to DN3 to achieve the first-ever conscious learning algorithm which is free from "deep learning" misconduct.[10] His research has been featured on 'Discovery Channel, Enel, and BBC.[11]
From 1983 to 1989, Weng's research work during his master's degree and Ph.D. degree focused on the analysis of the motion of objects and estimating 3D structures from motion.[12] He realized that such model-based approaches can provide with piecemeal insights but are too restrictive for understanding how animal brains learn vision and other brain skills. Soon after his Ph.D. degree work, he started Cresceptron.[13]
Cresceptron represented a direction that Weng later termed Autonomous Mental Development (AMD). In 1992, he and his collaborators pioneered the development of a framework titled Cresceptron for segmenting and recognizing real-world 3D objects from their images through automated learning.[13] The framework was tested for visual recognition, specifically recognizing 3-D objects from 2-D images and segmenting them from cluttered backgrounds without the need for handcrafted 3D models. It employed techniques such as stochastic distortion modeling, view-based interpolation, and a combination of individual and class-based learning approaches. Cresceptron achieved seven significant accomplishments, including the development of techniques like learning large-scale 3D objects with a deep convolutional neural network (CNN) and feature-independent learning for extensive datasets, among others. It was also established that Cresceptron significantly differs from later "deep learning" networks due to its approach of developing a sole network using Hebbian learning (i.e., unsupervised in all hidden layers).[14] [15]
Weng introduced another framework named SHOSLIF which provided a unified theory and methodology for comprehensive sensor-actuator learning.[16] It addressed single sensory problems as well as critical issues that Cresceptron faces, such as the automated selection of the most valuable features, the automatic organization of sensory and control information through a coarse-to-fine space partition tree, resulting in a remarkably low, logarithmic time complexity for content-based retrieval from extensive visual knowledge bases.[17] It also deals with handling invariance through learning, enabling online incremental learning, and facilitating autonomous learning, among other objectives.[18] [19]
From 1998 to 2010, Weng developed SAIL[20] and Dav robots[21] using sensory mapping models including self-aware self-effecting (SASE), staggered hierarchical mapping (SHM), and incremental hierarchical discriminant regression (IHDR) methods. It has been applied to the recognition of occluded objects,[22] speech recognition,[23] vision-guided navigation,[24] and range-based collision avoidance.[25]
Since 2005, Weng and his team have been working on the development of brain-like and cortex-like Developmental Networks (DNs) and their embodiments Where-What Networks (WWNs)[26] using brain-like architecture, including modeling pathways, laminar 6-layer cortex, and brain areas.[27] [28] In addition, they have analyzed how the brain deals with modulation, time, and space and have created three versions (DN1 through DN3) by 2023. A significant enhancement introduced in the transition from DN-2 to DN-3 involves initiating the brain-size network from a single-cell zygote. This means a fully autonomous process for brain patterning from a single cell. The key mechanisms of patterning include the Lobe Component Analysis (LCA)[29] and Synaptic Maintenance,[30] which automatically maintain the global smoothness of brain representation and local refinements of area representations. This approach enabled the developmental algorithm to progressively develop sensors, a complex brain, and motor functions in a sequential and self-organizing manner, ensuring that the wiring and pattern formation processes occur automatically from the initial conception stages throughout the entire life of the system.[31]
These Developmental Networks (DNs) and Where-What Networks (WWNs 1–9) have been developed for versatile visual learning in complex environments.[32] DNs can recognize objects and autonomously determine where and what to focus on using self-generated task context. Furthermore, these WWNs and DNs have been applied to general-purpose vision,[33] temporal visual event recognition,[34] vision-guided navigation,[35] learning audition while learning to speak,[36] and language acquisition as brain's responses to text temporal events.[37] Weng is the first to formally raise that robotic consciousness is necessary for AI, consciousness can and should be learned (i.e., developed), and proposed a fully implementable algorithm to do so. He proposed DN3[31] as the engine to conduct conscious learning[38] where a robot is able to become increasingly conscious, like an infant and then a child, through its 'living' experience in the physical world which typically include human parents and teachers. However, there is no central controller within DN3's skull, emphasizing that consciousness should not be statically handcrafted and must encompass elements beyond a programmer's design.[31]
Since 2016, Weng has alleged instances of plagiarism and post-selection misconduct on a worldwide scale, but the implicated institutions have not yet acknowledged his allegations.
Weng alleged that many deep learning networks that use images of 3D objects copied their key idea from Cresceptron[13] but almost all later deep learning publications did not cite Cresceptron. He highlighted that the Cresceptron (for 3D) is very different from the Neocognitron[39] (for 2D) because the Cresceptron is a fundamental departure from Neocognitron. Cresceptron enables a neural network to grow incrementally from a zero-neuron hierarchy and learn 3D objects from their 2D images in cluttered scenes. This is different from the aspect graphs of the 1990s and all other methods that had an inside-skull human teacher as a central controller.[40] This alleged plagiarism includes HMAX at MIT[41] and the ACM Turing Award 2018.[42] Without internal weight supervision like human manual selections[39] [41] and error-backprop,[42] feature learning and sharing in hidden areas of Cresceptron are based on (unsupervised) Hebbian mechanisms.[43]
Weng raised the issue of Post-Selection in AI and argued that it constitutes misconduct. He addressed that many AI methods require two steps in their training stage. The first step consists of training multiple systems by randomly fitting a fit data set. The second step consists of Post-Model Selection (Post-Selection). The Post-Selection chooses a few luckiest trained systems or relies on human manual parameter-tuning based on the systems’ errors on a validation data set. He alleged that Post-Selection in AI contains two types of misconduct: (1) cheating in the absence of a test, because the Post-Selection step belongs to the training stage; (2) hiding bad-looking data, because less lucky systems were not reported.[10] Weng further alleged that more categories of AI methods suffered from their Post-Selection steps, such as Neocognitron, HMAX, Deep Learning, Long Short-Term Memories, Extreme Learning Machines, Evolving Networks, Reservoir Computing, Transformers, Large Language Models, ChatGPT, and Bard, as long as they contain the Post-Selection step, which is either automatic or requires human manual tuning. He mathematically reasoned that the luckiest system on a validation set gives only an expected performance on a future test set that is only near the average performance of all trained systems on the validation set.[10]
Weng has sued institutions to address the issue of alleged misconduct outside of academia including Alphabet, in the United States District Court for the Western District of Michigan (Civil Action No. 1:22-cv-998)[44] and the US Court of Appeal 6th Circuit (Civil Action No, 23–1567).[45]