Sign in to view Steve’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Steve’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New York City Metropolitan Area
Sign in to view Steve’s full profile
Steve can introduce you to 5 people at Futurewei Technologies, Inc.
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
3K followers
500+ connections
Sign in to view Steve’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Steve
Steve can introduce you to 5 people at Futurewei Technologies, Inc.
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Steve
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Steve’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
- 14 AI…
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Activity
3K followers
-
Steve Tan liked thisSteve Tan liked thisThe Singularity has belonged exclusively to artificial minds, until now. For decades, whole-brain emulation has been the tantalizing counterpart to artificial intelligence: copy a biological brain, neuron by neuron and synapse by synapse, and run it. Today, for the first time, I am releasing a video from a company I helped found, Eon Systems PBC, demonstrating what we believe is the world's first embodiment of a whole-brain emulation that produces multiple behaviors. (Continue reading today's full edition ⬇️)
-
Steve Tan liked thisDeepSeek R1 is now available on Perplexity for free! Most importantly, the entire chain of thought is rendered for you to follow and verify. Perplexity now has verifiable sources, verifiable chain of thought, and verifiable answers as a result.Steve Tan liked thisDeepSeek R1 is now available on Perplexity to support deep web research. There's a new Pro Search reasoning mode selector, along with OpenAI o1, with transparent chain of thought into model's reasoning. DeepSeek on Perplexity is hosted in US/EU data centers - your data never leaves Western servers. The open source model is hosted completely independent of China. Your privacy and data security is our priority. We're increasing the number of daily uses for both free and paid as add more capacity during the day. Stay tuned! #deepseek #ai #perplexity #opensource
-
Steve Tan liked thisSteve Tan liked thisAnnouncing our exciting new resident program at Perplexity. No matter where you are currently in your career or which industry you've had experience in, Perplexity offers a unique experience to learn and ship product. First class scheduled for Feb 2025 and second class scheduled for June 2025. Stay curious and keep building! https://lnkd.in/edWNAESu
-
Steve Tan liked thisWe need to all be aware of the growing racism in our country against Chinese immigrants. Attributing one student doctoring their data to an entire ethnicity or race is abhorrent behavior for anyone, much less for a well-respected or formerly well-respected MIT professor. I highly doubt that this student said this quote and it was likely made up by the professor herself. This leads to racial profiling and witch hunts such as the China Initiative that ruined the lives of dozens of professors and the death of one by suicide. This type of rhetoric is what incites distrust in communities and ultimately hate crimes. Former Stanford President Marc Tessier-Lavigne was forced to resign after it was revealed that he had manipulated and falsified data in 12 of his research reports. Would Professor Picard dare insinuate that white American men have no morals or values? Or did she think that Asians are easy targets that wouldn’t call her out on her blatant racism? If she is one of the preeminent minds and thought leaders in AI and is openly biased and racist, how can we expect our AI models to not share and perpetuate that dangerous bias? Thank you for bringing this to my attention Jamaal Glenn.Steve Tan liked thisThe greatest irony at #NeurIPS2024 was witnessing an MIT professor Rosalind Picard, a self-proclaimed AI ethics expert, openly discriminate against Chinese researchers during a keynote speech. This blatant display of racism is utterly unacceptable and directly contradicts the principles of equity, inclusion, and respect that the global research community—and especially the NeurIPS community—strives to uphold. Ethical challenges are universal and should never be attributed to any nationality or culture. As researchers, we carry a collective responsibility to ensure our words and actions embody the values of the inclusive world we seek to create. If we, as humans, fail to align ourselves with these values, how can we ever hope to build responsible AI that advances humanity? We urge #NeurIPS and #MIT to thoroughly investigate this incident and take the necessary steps to ensure accountability. Racism has no place in science—now or ever. #StopRacism #InclusionInAI #RespectInResearch #EthicsInAI #NeurIPS2024 #AIEthics #AI4Humanity
Experience & Education
-
Futurewei Technologies, Inc.
********* ********
-
******* ***
******** *** * ***
-
******
******** *********
-
******* **********
****** ** ********** * *** ************* *************** ********** **** ******** undefined
-
View Steve’s full experience
See their title, tenure and more.
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View Steve’s full profile
-
See who you know in common
-
Get introduced
-
Contact Steve directly
Other similar profiles
Explore more posts
-
LavX Managed Systems
76 followers
New research underscores a paradigm shift in performance optimization: CPU speeds now vastly outpace memory access, redirecting focus from instruction-level tweaks to data structure design. While compilers master low-level code, they falter on data layout transformations, elevating early architectural choices as critical for enterprise-scale systems. For AI/ML leaders in SaaS and fintech, this demands data-centric engineering to unlock ROI through superior efficiency and scalability. #EnterpriseAI #DigitalTransformation #TechLeadership
1 Comment -
Mahesh Godavarti
Automate.ai • 1K followers
MeanFlow: Residual Learning for One-Step Image Generation Before the introduction of skip-connections in ResNets (2015), deep neural networks often struggled with optimization as depth increased. Training directly to map x→y was harder than learning the residual x→(y−x) and adding x back at the end. This residual formulation simplified the learning task and made deeper architectures practical. A similar situation has existed in image generation. Diffusion models can produce high-quality images from noise, but typically require dozens or hundreds of iterative refinement steps. One-step generation — mapping noise directly to an image — is possible, but has generally not matched the quality of multi-step approaches. The recent MeanFlow paper reframes this problem in a way that parallels residual learning. Instead of predicting the entire image from noise, it learns the average change required to transform noise into the final image — the analogue of a residual in this context. This “average velocity” is smoother and more structured than the raw image, making it easier to learn. Applying this change in one step produces results that close much of the quality gap with multi-step methods. In short, MeanFlow does for one-step generative modeling what skip-connections did for deep classification networks: it turns a complex direct mapping into a simpler residual-style prediction problem. #DeepLearning #GenerativeAI #ImageGeneration #NeuralNetworks #DiffusionModels #MachineLearning #MeanFlow #ResidualLearning #ResNet Paper: https://lnkd.in/g-74xvTK
5
-
Quantum Zeitgeist
17K followers
Neutral-atom Quantum Processors Achieve 84% Feature Selection Accuracy with 2.3% Redundancy Reduction Researchers demonstrate that programmable arrays of neutral atoms successfully identify the most important features in datasets, improving the accuracy of machine learning models while dramatically reducing the number of features required for effective classification. #quantum #quantumcomputing #technology https://lnkd.in/eEWiQtQ8
1
-
Quantum Zeitgeist
17K followers
Neutral-atom Quantum Processors Achieve 84% Feature Selection Accuracy with 2.3% Redundancy Reduction Researchers demonstrate that programmable arrays of neutral atoms successfully identify the most important features in datasets, improving the accuracy of machine learning models while dramatically reducing the number of features required for effective classification. #quantum #quantumcomputing #technology https://lnkd.in/e5EN-uDM
1
-
Quantum Zeitgeist
17K followers
Neutral-atom Quantum Processors Achieve 84% Feature Selection Accuracy with 2.3% Redundancy Reduction Researchers demonstrate that programmable arrays of neutral atoms successfully identify the most important features in datasets, improving the accuracy of machine learning models while dramatically reducing the number of features required for effective classification. #quantum #quantumcomputing #technology https://lnkd.in/eEWiQtQ8
-
JuliaHub
35K followers
Traditional causal modeling defines explicit signal flow from inputs to outputs—but as systems grow, it becomes rigid and hard to extend. Acausal modeling focuses on the physical relationships between components, letting equations form automatically. In this post, see how Dyad, built on Julia and #SciML, makes acausal #modeling scalable, reusable, and ideal for complex system design. https://lnkd.in/eTCEdqUZ #JuliaLang #SystemSimulation #ModelBasedDesign #Dyad
8
-
online forms
Jarf Pardazan Mehr • 1K followers
VCRL: Variance-based Curriculum Reinforcement Learning for Large Language Models Curriculum Learning in Reinforcement Learning The article discusses a novel curriculum learning framework for reinforcement learning, called VCRL, which utilizes variance in rollout groups to gauge sample difficulty. This approach selects high-variance samples and employs a replay memory mechanism to stabilize training. VCRL is compared to existing rollout-based methods, such as GRPO, DAPO, and GSPO, highlighting its unique features and potential benefits. The core idea of VCRL revolves around the concept of difficulty in reinforcement learning, emphasizing the importance of sampling strategies and memory mechanisms in improving training efficiency. Critical Evaluation The strengths of VCRL include its clear motivation, concrete sampling strategy, and comprehensive benchmark evaluation. However, weaknesses such as the reliance on heuristics without theoretical backing, potential limitations in sparse reward contexts, and lack of discussion on bias toward mid-difficulty samples are notable concerns. The exposition of VCRL is concise, showcasing its novelty in removing value models, but notation density, absent proofs of variance efficacy, and unaddressed computational overhead of multiple rollouts limit immediate practical assessment. Key terms like reinforcement learning, curriculum learning, and variance-based sampling are crucial in understanding the article's main points. In terms of biases, the article may exhibit a potential bias toward mid-difficulty samples, which could impact the generalizability of the results. The implications of VCRL are significant, as it offers a new perspective on curriculum learning in reinforcement learning, highlighting the importance of sample difficulty and memory mechanisms. However, further research is needed to address the limitations and concerns raised in the article, particularly in terms of theoretical backing, computational overhead, and robustness across tasks. The concept of difficulty in reinforcement learning is a critical aspect of VCRL, and its implications for reinforcement learning and curriculum learning are substantial. Conclusion and Call to Action The article provides a valuable contribution to the field of reinforcement learning, highlighting the potential benefits of curriculum learning and variance-based sampling. The implications of VCRL are significant, and its potential impact on the field of reinforcement learning is substantial. However, further research is needed to address the limitations and concerns raised in the article. What are your thoughts on the potential of VCRL in shaping the future of reinforcement learning and curriculum learning? How do you see this research impacting your industry or work? arXiv page: https://lnkd.in/edPbU-4Y
-
Bayes Labs
3K followers
Research Paper Highlights: "Evaluating Large Language Models in Scientific Discovery", By Zhangde Song et al. Large language models are increasingly used in scientific research, but existing benchmarks fail to capture how real scientific discovery unfolds. Most evaluations focus on isolated factual questions, ignoring hypothesis formation, experimentation, and iterative reasoning. This work introduces a scenario-grounded Scientific Discovery Evaluation (SDE) framework to assess LLMs on realistic, expert-defined research workflows. Challenges: - Decontextualized evaluation: Traditional benchmarks overlook the project-level reasoning central to scientific discovery. - Misleading scaling signals: Gains from larger models and stronger reasoning do not translate proportionally to discovery tasks. - Hidden weaknesses: Existing benchmarks mask systematic reasoning gaps that emerge only in end-to-end research scenarios. Key Takeaways: - Scenario-grounded design: Expert-defined research projects are decomposed into modular scenarios across biology, chemistry, materials, and physics. - Two-level evaluation: Models are assessed on both question-level accuracy and project-level scientific reasoning, including hypothesis generation and interpretation. - Revealing insights: Top-tier LLMs show consistent performance gaps, large variability across scenarios, and no clear winner across discovery projects. This research emphasizes that true scientific intelligence cannot be measured by isolated QA performance alone. The SDE framework exposes fundamental limitations of current LLMs in discovery-driven reasoning, showing they remain far from scientific superintelligence. At the same time, it highlights LLMs’ emerging potential in guided exploration and provides a reproducible path forward for advancing AI-assisted scientific discovery. Further reading-https://lnkd.in/gYEpMaDB
5
-
Abdulkadir Celik
University of Southampton • 7K followers
📣 #CCSL's position paper on #NearField-#ISAC has been accepted for publication in IEEE Wireless Communications Magazine. ⁉️Ultra-massive #MIMO systems operating within the NF region utilize spherical wavefronts, enabling precise beam manipulation in both angular and distance domains through #BeamFocusing. While this approach dramatically boosts communication capacity and sensing precision, it also introduces increased beam management complexity. 💡🎯 We propose a 3️⃣-Step interplay between communication codebooks and Space-Time Adaptive Processing (#STAP) to reduce overall beam training overhead and sensing complexity: 1️⃣ We begin by leveraging #FarField (FF) discrete Fourier transform (#DFT) codebooks, periodically swept during synchronization signal block (#SSB) transmissions by #5G networks. NF users observing DFT beams with spherical wavefronts experience an angular spread dependent on their angle and range, allowing them to reference a look-up table for coarse estimations. 2️⃣ The initial range and angle estimates are further refined using #polar #codebooks, which oversample beams within the candidate polar regions, enhancing estimation accuracy and reducing training overhead. 3️⃣ Utilizing the refined estimates, we reduce the dimensions of the #radar #datacube, enabling a low-complexity Space-Time Adaptive Processing (#STAP) execution. This results in a three-orders-of-magnitude reduction in computational complexity, providing highly precise estimates of range, angle, radial velocity, and transverse velocity (unattainable with planar wavefronts in the far-field) 🔍 This integrated approach significantly minimizes the overhead associated with beam-training in both angular and distance domains while enhancing estimation accuracy. By co-designing NF communication and sensing systems, our framework ensures high efficiency and seamless integration of both functionalities, paving the way for more intelligent and efficient future wireless networks. 👥 Ahmed Hussain, Asmaa Abdallah, Ph.D,Abdulkadir Çelik, Ahmed Eltawil 🏛 KAUST (King Abdullah University of Science and Technology) KAUST CEMSE 🔗ResearchGate: https://lnkd.in/eAbuHZjc IEEE IEEE Signal Processing Society IEEE Computer Society IEEE Xplore IEEE ComSoc Young Professionals ACM, Association for Computing Machinery ACM arXiv Machine Learning Research at arXiv IEEE Communications Society IEEE Access IEEE The Institute Proceedings of the IEEE Computer Engineering Computer Science Computer Networking Computer Literacy Shannon Wireless Huawei Ericsson 6G-life Wireless Networking Wireless Communications Systems Huawei Wireless 6G Flagship telecomhall International Telecommunication Union
95
1 Comment -
Praveen Kumar Pokala, PhD
JPMorganChase • 13K followers
Memory Is Not Storage In Agentic AI ----------------------------------------- Hallucinations are often blamed on model uncertainty, but in agentic systems they are more often memory pathologies. Intelligence is not about how much a system remembers, but what it remembers, what it forgets, and when experience becomes belief. Short-term memory preserves coherence, episodic memory preserves honesty, and long-term memory preserves identity. Confuse them and the system becomes confidently wrong; separate them and learning becomes possible without drift. Memory is not storage, it is a theory of cognition. Models generate possibilities, decoding chooses actions, but memory determines who the system becomes over time. For a deeper cognitive view of this, see my blog on agentic memory and learning (https://lnkd.in/gmmreWqf). #AgenticAI #AIMemory #AIArchitecture #AIHallucinations #CognitiveAI
20
-
KYUNGJUN LIM
freederia • 5K followers
New Post: Reinforcement‑Learning‑Guided Thermodynamic Protocols for Enhancing Ergodic Fluctuation Relations in Quantum Spin Chains - https://lnkd.in/giej7PcG ### Abstract We introduce a data‑driven framework that learns optimal, time‑dependent control fields for a finite quantum spin‑½ chain to maximize the average entropy production in a nonequilibrium protocol. By integrating the Lindblad master equation with a deep deterministic policy gradient \(DDPG\) agent, the method automatically discovers driving schedules that surpass conventional deterministic protocols \[…\]
-
Christophe Pere, PhD
Nord Quantique • 24K followers
> Sharing Resource < Survey on Quantum Federated Learning by Dinh C. Nguyen, Md Raihan Uddin, Shaba Shaon, Ratun Rahman, Octavia Dobre, Dusit Niyato Abstract: Quantum federated learning (QFL) is a combination of distributed quantum computing and federated machine learning, integrating the strengths of both to enable privacy-preserving decentralized learning with quantum-enhanced capabilities. It appears as a promising approach for addressing challenges in efficient and secure model training across distributed quantum systems. This paper presents a comprehensive survey on QFL, exploring its key concepts, fundamentals, applications, and emerging challenges in this rapidly developing field. Specifically, we begin with an introduction to the recent advancements of QFL, followed by discussion on its market opportunity and background knowledge. We then discuss the motivation behind the integration of quantum computing and federated learning, highlighting its working principle. Moreover, we review the fundamentals of QFL and its taxonomy. Particularly, we explore federation architecture, networking topology, communication schemes, optimization techniques, and security mechanisms within QFL frameworks. Furthermore, we investigate applications of QFL across several domains which include vehicular networks, healthcare networks, satellite networks, metaverse, and network security. Additionally, we analyze frameworks and platforms related to QFL, delving into its prototype implementations, and provide a detailed case study. Key insights and lessons learned from this review of QFL are also highlighted. We complete the survey by identifying current challenges and outlining potential avenues for future research in this rapidly advancing field. Link: https://lnkd.in/eRen6hgu #quantummachinelearning #quantumfederatedlearning #quantumalgorithms #research #paper
36
-
Yan Barros
BiaTech Corporation • 9K followers
Physics-Informed Neural ODEs with Scale-Aware Residuals for Learning Stiff Biophysical Dynamics Kamalpreet Singh Kainth, Prathamesh Dinesh Joshi, Raj Abhijit Dandekar, Rajat Dandekar, Sreedat Panat https://lnkd.in/dtRD_ide This paper tackles a critical challenge in using Neural ODEs for modeling stiff biophysical systems, specifically the difficulty in achieving stable and accurate long-term predictions. The core idea is to improve training stability and accuracy by incorporating physics-informed regularization with scale-aware residual normalization within a Neural ODE framework. The authors propose "PI-NODE-SR," which combines two key elements: 1. Physics-Informed Regularization: They leverage the known governing equations (e.g., Hodgkin-Huxley) to define a physics-informed loss term that penalizes deviations from the expected behavior. This is a standard PINN approach, but the key is how they handle the residual. 2. Scale-Aware Residual Normalization: This is the novel contribution. Stiff systems often involve state variables evolving on vastly different timescales. Directly applying a physics-informed loss can lead to one timescale dominating the training process, hindering convergence and accuracy. To address this, they normalize the residual associated with each state variable by a scaling factor. This factor is chosen to balance the contributions of each variable to the overall loss, preventing the faster dynamics from overwhelming the slower ones. They use a low-order explicit solver (Heun method) to calculate the residual. The authors demonstrate the effectiveness of PI-NODE-SR on the Hodgkin-Huxley equations, showing that it can learn from a single oscillation and extrapolate accurately over longer time horizons. Importantly, they show that the method can recover morphological features in the gating variables that are typically only captured by higher-order solvers. ----- This work directly addresses a significant limitation of Neural ODEs in scientific applications: their struggle with stiff systems. By introducing scale-aware residual normalization, the authors provide a principled way to stabilize training and improve the accuracy of long-term predictions. This is particularly relevant for: - Biophysical modeling: Accurately simulating neuronal dynamics, cardiac electrophysiology, and other complex biological processes. - Chemical kinetics: Modeling reaction networks with widely varying reaction rates. - Multi-physics simulations: Situations where different physical processes evolve on different timescales. The ability to use lower-order solvers with neural correction to achieve accuracy comparable to higher-order methods has significant implications for computational efficiency. While the method's sensitivity to initialization is noted, the overall approach offers a promising direction for developing more robust and efficient Physics-Informed Neural ODEs for a wide range of scientific applications.
146
1 Comment -
Valeria Gonzalez Perez
UC San Diego • 2K followers
𝑮𝒖𝒂𝒓𝒅𝒊𝒂𝒏-𝒓𝒆𝒈𝒖𝒍𝒂𝒓𝒊𝒛𝒆𝒅 𝑺𝒂𝒇𝒆 𝑶𝒇𝒇𝒍𝒊𝒏𝒆 𝑹𝒆𝒊𝒏𝒇𝒐𝒓𝒄𝒆𝒎𝒆𝒏𝒕 𝑳𝒆𝒂𝒓𝒏𝒊𝒏𝒈 𝒇𝒐𝒓 𝑺𝒎𝒂𝒓𝒕 𝑾𝒆𝒂𝒏𝒊𝒏𝒈 𝒐𝒇 𝑴𝒆𝒄𝒉𝒂𝒏𝒊𝒄𝒂𝒍 𝑪𝒊𝒓𝒄𝒖𝒍𝒂𝒕𝒐𝒓𝒚 𝑫𝒆𝒗𝒊𝒄𝒆𝒔🫀🩺 Presented at NeurIPS by UCSD PhD CSE student, Ayşın Tümay, this research offers a great approach in the field of MedicalAI by assisting clinicians through AI powered decision making🥼 In other words, this research supports clinicians with an automated decision making strategy for gradually reducing dependence on Mechanical Circulatory (MCS) Devices, which act as a temporary blood flow support for patients who suffer from heart failure or cardiogenic shock, by utilizing historical patient data (offline) rather than real time in the moment (online) patient data - prioritizing safety for patient care in such a delicate weaning process. Some things to note is that in order to ensure safety, the Reinforcement Learning algorithm penalizes suggested actions which are highly disparate from its training data (Out Of Distribution Suppression) + expert clinical knowledge was incorporated in order to shape the AI's goal system towards specific medical desired outcomes (maintaining the patient stable through the transition off the medical device). For in depth technical details, author Ayşın Tümay has shared the article below! As someone who is deeply interested in AI for Medicine, thank you Aysin for sharing your work and for being a great resource! p.s I would definitely reach out if you are someone who is curious about AI for Cardiology👀
4
-
PyTorch
318K followers
Our latest PyTorch community blog from the SGLang Team covers how SGLang advances hybrid Mamba plus Attention models with improved memory efficiency, prefix caching, speculative decoding, and serving performance. 📖 Read the blog: https://lnkd.in/gfDZkZWF #PyTorch #AIInfrastructure #SGLang #LLM
110
1 Comment -
Stefano Nichele
Østfold University College… • 2K followers
Our book chapter "Studying Encoder–Decoder Relation Between Cellular Automata to Uncover Their Computational Structure" is finally published in the book "Advances in Cellular Automata, Volume 2". This work was led by Barbora Hudcova, in collaboration with Tomas Mikolov and myself. Link: https://lnkd.in/d6YeTD4m Preprint: https://lnkd.in/dpiVisDi We use encoders-decoders to simulate elementary CA, and we identify a rich hierarchy of their relationships that were not known before. We show that the number of elementary automata with unique dynamics can be reduced from 88 to 52!
52
1 Comment
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content