Junzhe Zhang (@junzhezhang12) 's Twitter Profile
Junzhe Zhang

@junzhezhang12

Assistant Professor of EECS @SyracuseU; Phd at @Columbia; Interested in Causal Inference, RL, and AI.

ID: 1531684866722504707

linkhttps://junzhez.com calendar_today31-05-2022 17:12:19

39 Tweet

49 Followers

20 Following

Junzhe Zhang (@junzhezhang12) 's Twitter Profile Photo

If we view physics as an endeavor to find the most efficient compression to summarize real-world data, the line between fields starts to blur.

Junzhe Zhang (@junzhezhang12) 's Twitter Profile Photo

Please check out our work "Causal Imitation for Markov Decision Processes: a Partial Identification Approach" (with Kangrui Ruan, Sharon (Colubmia) and Elias Bareinboim) at #NeurIPS2024! - Thu 12 Dec 4:30 pm — 7:30 pm - West Ballroom A-D #5002 - Link: neurips.cc/virtual/2024/p…

Please check out our work "Causal Imitation for Markov Decision Processes: a Partial Identification Approach" (with <a href="/realDarrenRuan/">Kangrui Ruan</a>, <a href="/Sharon64803136/">Sharon (Colubmia)</a> and <a href="/eliasbareinboim/">Elias Bareinboim</a>) at #NeurIPS2024!
- Thu 12 Dec 4:30 pm — 7:30 pm
- West Ballroom A-D #5002
- Link: neurips.cc/virtual/2024/p…
Elias Bareinboim (@eliasbareinboim) 's Twitter Profile Photo

2/5 "Automatic Reward Shaping from Confounded Offline Data" (w/Mingxuan Li, Junzhe Zhang) Wed, 1 PM (East Ballroom, 1802) Link: causalai.net/r123.pdf Reward shaping is a popular technique for addressing the sparse reward problem in RL by injecting additional

2/5  "Automatic Reward Shaping from Confounded Offline Data"
 (w/<a href="/Mingxuan0422/">Mingxuan Li</a>, <a href="/JunzheZhang12/">Junzhe Zhang</a>)
 
  Wed, 1 PM (East Ballroom, 1802)
 
  Link: causalai.net/r123.pdf

Reward shaping is a popular technique for addressing the sparse reward problem in RL by injecting additional
Elias Bareinboim (@eliasbareinboim) 's Twitter Profile Photo

4/5 (UAI) "Eligibility Traces for Confounding Robust Off-Policy Evaluation" (w/ Junzhe Zhang) Link: causalai.net/r105.pdf The main goal of off-policy evaluation is to enable agents to learn from observational data without deploying new policies. While current methods

4/5 (UAI) "Eligibility Traces for Confounding Robust Off-Policy Evaluation" (w/ <a href="/JunzheZhang12/">Junzhe Zhang</a>)
 
 Link: causalai.net/r105.pdf

The main goal of off-policy evaluation is to enable agents to learn from observational data without deploying new policies. While current methods
Elias Bareinboim (@eliasbareinboim) 's Twitter Profile Photo

5/5 Last but definitely not least, I’m honored to be giving a keynote on Wednesday (7/23) titled "Towards Causal Artificial Intelligence." For details, see: auai.org/uai2025/keynot… Here’s a short abstract: While many AI scientists and engineers believe we are on the verge of

5/5 Last but definitely not least, I’m honored to be giving a keynote on Wednesday (7/23) titled "Towards Causal Artificial Intelligence." For details, see: auai.org/uai2025/keynot…

Here’s a short abstract:

While many AI scientists and engineers believe we are on the verge of
Junzhe Zhang (@junzhezhang12) 's Twitter Profile Photo

ACs implementing the review protocols can frustrate some reviewers. You can literally read it through their replies and comments. Lol #NeurIPS

Junzhe Zhang (@junzhezhang12) 's Twitter Profile Photo

The sooner we recognize that LLMs' "implicit reasoning" is simply MCMC sampling or the Euler method approximating a continuous function, the easier it becomes to make progress on its inference.