slide1 n.
Download
Skip this Video
Loading SlideShow in 5 Seconds..
ATLAS@LHC PowerPoint Presentation
Download Presentation
ATLAS@LHC

Loading in 2 Seconds...

play fullscreen
1 / 34

ATLAS@LHC - PowerPoint PPT Presentation


  • 147 Views
  • Uploaded on

Exp High Energy Physics in Lund fusion in 2003 between Particle Physics and the High Energy Heavy Ion group Div head: Paula Eerola. ALICE@LHC. Electronics. Overview of div. Human interations. p+p. e + e. Pb+Pb. e + p. M Curie EST @Lund. PHENIX@RHIC. H1@HERA. TPC@ILC. ATLAS@LHC.

loader
I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
capcha
Download Presentation

PowerPoint Slideshow about 'ATLAS@LHC' - payton


An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.


- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript
slide1

Exp High Energy Physics in Lundfusion in 2003 between Particle Physics and the High Energy Heavy Ion groupDiv head: Paula Eerola

slide2

ALICE@LHC

Electronics

Overview of div.

Human

interations

p+p

e + e

Pb+Pb

e + p

M Curie EST

@Lund

PHENIX@RHIC

H1@HERA

TPC@ILC

ATLAS@LHC

SUSY, Higgs,...

CP-violation

Detector

developm

Problem

Based

Learning

Proton structure,...

Quark-gluon plasma,...

Grid

people and projects
People and projects
  • ATLAS (pp @LHC)
    • P. Eerola, V. Hedberg, G. Jarlskog (emeritus), O. Smirnova, T. Åkesson.
    • PhD students N.Boelert, J.Groth-Jensen, J. Weina
    • Guests: S.Seidel, E.Gazis
  • GRID (distributed data handling)
    • P. Eerola, B. Kónya, O. Smirnova, J. Jönemo, PhD student: P. Gros (50%)
  • ALICE (Heavy ions @LHC)
    • P. Christiansen (fo.ass), H.-Å. Gustafsson, A. Oskarsson, I. Otterlund (emeritus), E. Stenlund PhD students: P. Gros (50%), A. Dobrin
  • PHENIX (Heavy Ions @ RHIC)
    • P. Christiansen, H.-Å. Gustafsson, A. Oskarsson, E. Stenlund.
    • PhD students: E. Haslum, S. Rosendahl
  • H1 (ep @ HERA)
    • L. Jönsson.
    • PhD students: M. Hansson, S. Osman.
  • ILC-TPC (detector development)
    • L. Jönsson, V. Hedberg, A. Oskarsson, E.Stenlund.

Engineers: B.Lundberg,U.Mjörnmark, L. Österman

the atlas experiment
The ATLAS experiment

ATLAS

Lund Status by Vincent Hedberg tomorrow

hep est teachers view
HEP EST, teachers view

--7EU+1LU students started october 2006

--4 exp, 3+1 theory

--7 countries: Belgium,China 2, Denmark, England, Italy, Romania, Sweden

--40p courses, finish in november

--all courses use PBL methodology

Exp Methods in HEP, Statistics and data analysis, QFT, numerical methods, particle phenomenology, SM and extension

--a lot of work

--teachers think PBL was fun

--8 PhD students will visit for 6 months

--Alex Dobrin gives students view tomorrow

phenix at rhic
PHENIX at RHIC
  • Au+Au at 100 + 100 GeV per nucleon
  • Polarised pp at 100 + 100 GeV (250+250 )
  • Taking data since 2001
total data volumes
Total Data Volumes

RHIC

  • Plot courtesy of Tom Throwe (RCF)

3 months exp

publication summary
Publication Summary

RHIC

  • Since 2001:
    • 44 PRL’s
    • 15 Phys. Rev. C
    • 4 Phys. Rev. D
    • 2 Phys. Lett. B
    • 1 Nucl. Phys. A (White Paper)
  • ~ 3500 citations
  • Most-cited paper from RHIC:
    • “Suppression of hadrons with large transverse momentum in central Au+Au collisions at √sNN= 130 GeV ”,K. Adcox et al. , Phys.Rev.Lett. 88:022301 (2002), nucl-ex/0109003
  • 12 other papers with > 100 citations

5 PHENIX PhDs from Lund

David Silvermyr, staff pos @ ORNL, PHENIX/ALICE

Pål Nilsson, staff pos @ Texas U., ATLAS

Henric Tydesjö, CERN Fellow, ALICE

Still in business

Up to 2006

slide12

5000 cards made

in Lund

TPC before placed in magnet

first cosmic ray data
First Cosmic Ray Data

OROC

IROC

Side view

3-dimensional view of a shower induced by cosmic rays

startup plan quite tentative
Startup plan(quite tentative)

December 07: 1 month Cosmic running

January 08: further Cosmics <1 month

Feb-March 08: Install remaining detectors

1May 08: Close magnet doors, wait for pp collisions full energy. Alice best for low Pt data complementary to ATLAS/CMS

End of 08: Possibly short PbPb, But may depend on pp status

April 09: long pp

Nov-dec 09: month PbPb

h1 at hera
H1 at HERA
  • H1 dismantled this summer. 15 years of data
  • Albert Knutsson, PhD march 07. Now postdoc at DESY
    • Forward jet production in deep inelastic scattering at HERA
    • Conclusions: First evidence for new parton dynamics, that can not be described by the resolved photon model.
  • Magnus Hansson, PhD defence sept 28-07.
    • Azimuthal correlations in dijet events from deep inelastic positron-proton scattering at HERA
    • Conclusions: Measurement directly sensitive to the amount of transverse momentum emitted in the parton cascade. Deviations to predictions by NLO calculations and QCD models are observed. A fit of the unintegrated gluon density has been performed, which for the first time gives a constraint of the intrinsic transverse momentum of the gluon propagator.
  • Data analysis continues by Sakar Osman. PhD mid next year
    • A study of minijets in deep inelastic electron proton scattering at HERA
grid activities in lund
Grid activities in Lund

Projects:

EU KnowARC: develops new Grid middleware; Lund provides the Project Leader (B.Kónya)

Design and early working prototypes are ready

NDGF: NOS-N funded, Nordic Grid Infrastructure, Tier1 for LHC computing; Lund provides the CERN Coordinator (O.Smirnova)

Tier1 is set up and running, participates in ALICE and ATLAS production and tests

NGIn: Nordunet3 project; focus on grid education and applications

4 PhD students in Nordic countries (P.Gros in Lund)

NorduGrid: a collaboration that initiated the projects above

People:

Paula Eerola – NorduGrid, KnowARC, NDGF

Philippe Gros – Ph.D. student (NGIn – ALICE applications)

Johan Jönemo – KnowARC

Balázs Kónya – NorduGrid, KnowARC

Ulf Mjörnmark – NorduGrid, KnowARC

Оxana Smirnova – NDGF, NorduGrid, KnowARC

Short-term visitors, master students

9/18/2014

16

detector development for ilc international linear collider
Detector development for ILC (international linear collider)

Joint effort of Lund HEP division

L. Jönsson, V. Hedberg, B. Lundberg, U.Mjörnmark, A.Oskarsson, E. Stenlund, L. Österman

EUDET, FP6 III project to position Europé in detector for ILC

+ VR counterfinancing

EUDET partners: CERN, DESY, NIKHEF, CEA, CNRS,CSIC,MPI,INFN

+ universities

EUDET:

test beamline at DESY

high resolution tracking

High resolution calorimetry

ilc timeline

2012

start constr

2019

end constr

2005 2006 2007 2008 2009 2010

ILC Timeline

Global Design Effort

Project

Baseline configuration

Reference Design

Technical Design

ILC R&D Program

Expression of Interest to Host

International Mgmt

baseline configuration schematic
500 GeV CMBaseline Configuration - Schematic
  • 1 TeV CM

~30km, 31.5MV/m

2820 bunches, spaced 300ns 5 times per second

Luminosity 2 1034

the detector order of magnitude better resolution
The detector. Order of magnitude better resolution.

TPC with GEM readout chambers

1*6mm pads, 5M channels

slide21

Lund/CERN readout electronics and DAQ task

Within half a year:

-10000 channel system based on ALICE readout electronics

-new amplifier (Nov 2007)

programmable: polarity, gain, shaping

-ALTRO digitizer chip with 40MHz sampling

Longer term:

Full integration of amplifier and digitizer for final TPC

slide23

Digital

Circuit

Need new preamp-shaper chip,

Programmable pol. Gain, shaping time

190nm process

ALICE front end card

FEC (Front End Card) - 128 CHANNELS

(CLOSE TO THE READOUT PLANE)

DETECTOR

Power

consumption:

< 40 mW / channel

L1: 6.5ms

1 KHz

8 CHIPS

x

16 CH / CHIP

8 CHIPS

x

16 CH / CHIP

drift region

88ms

L2: < 100 ms

200 Hz

ALTRO

gating grid

PASA

ADC

RAM

anode

wire

DDL

(3200 CH / DDL)

CUSTOM IC

(CMOS 0.35mm)

pad

plane

570132 PADS

CUSTOM IC (CMOS 0.25mm )

  • GAIN EQUALIZ.
  • LINEARIZATION
  • BASELINE CORR.
  • TAIL CANCELL.
  • ZERO SUPPR.

CSA

SEMI-GAUSS. SHAPER

1 MIP = 4.8 fC

S/N = 30 : 1

DYNAMIC = 30 MIP

10 BIT

< 12 MHz

MULTI-EVENT

MEMORY

GAIN = 12 mV / fC

FWHM = 190 ns

FEE FOR THE NA49 AND STAR TPCs

the physics agenda for the ilc
Higgs

- The Standard Model Higgs

- SUSY Higgs

Non-SUSY extensions of SM

SUSY

- Minimal Supersymmetric Standard Model (MSSM)

- The Minimal Supergravity model (mSUGRA)

- Gauge-Mediated SUSY Breaking (GSMB)

- Anomaly-Mediated SUSY Breaking (AMSB)

Alternative theories

- Extra Dimensions

- Strong electroweak symmetry breaking

- Compositness

Precision measurements

- Electroweak Gauge bosons

- Extended Gauge theories

- Top quark physics

- Quantum Chromodynamics

The physics agenda for the ILC

 J.A. Aguilar-Saavedra et al., hep-

ph/0106315

 T. Abe et al., hep-ex/0106055

 K. Abe et al., hep-ph/0109166

 G. Weiglein et al., hep-ph/0410364

Very much the same as LHC

Why ILC?

Complementarity to LHC

  • Clean entrance channel
  • Matched by precision measurements
tpc alice

5.6 m

TPC ALICE

GAS VOLUME

88 m3

DRIFT GAS

90% Ne - 10%CO2

Co2 insulation

E

E

E

E

Drift volume

E

88ms

400 V / cm

1.6

P

b

P

NE / CO2

88ms

510 cm

E

Central electrode

Readout plane segmentation

18 trapezoidal sectors

each covering 20 degrees in azimuth

End plate

5 m

the baseline machine 500gev
The Baseline Machine (500GeV)

~30 km

ML ~11.2km (G = 31.5MV/m)

20mr

RTML ~1.6km

BDS 5km

2mr

e+ undulator @ 150 GeV (~1.2km)

R = 955m

E = 5 GeV

x2

not to scale

the ilc accelerator
The ILC Accelerator
  • 2nd generation electron-positron Linear Collider
  • Parameter specification
    • Ecms adjustable from 200 – 500 GeV
    • Luminosity ∫Ldt = 500 fb-1 in 4 years
    • Ability to scan between 200 and 500 GeV
    • Energy stability and precision below 0.1%
    • Electron polarization of at least 80%
    • Options for electron-electron and g-g collisions
    • The machine must be upgradeable to 1 TeV
  • Three big challenges: energy, luminosity, and cost
slide30

Scope of the 500 GeV machine

Main linacs length ~ 21 km, 16,000 RF cavities (total)

RF power ~ 640 10-MW klystrons and modulators (total)

Cryoplants ~ 11 plants, cooling power 24 kW (@4K) each

Beam delivery length ~ 5 km, ~ 500 magnets (per IR)

Damping ring circumference ~ 6.6 km, ~400 magnets each

Beam power ~ 22 MW total

Site power ~ 200 MW total

Site footprint length ~ 47 km (for future upgrade > 1 TeV)

Bunch profile at IP ~ 500 x 6 nm, 300 microns long

slide31

Elements of the BCD

  • Parameter plane established
    • TESLA designed for 3.4e34 but had a very narrow operating range
    • ILC luminosity of 2e34 over a wide range of operating parameters
      • Bunch length between 500 and 150 um
      • Bunch charge between 2e10 and 1e10
      • Number of bunches between ~1000 and ~6000
      • Beam power between ~5 and 11 MW
  • Superconducting linac at 31.5 MV/m
    • Cavities qualified at 35 MV/m in vertical tests
    • Expect an average gradient of 31.5 MV/m to be achieved
      • Rf system must be able to support 35 MV/m cryomodules
    • This still requires extensive R&D on cavities and rf sources
slide32

5GeV

Final ILC design?

~ 1 TeV

8-10 years away,

somewhere

slide33

GAS VOLUME

88 m3

DRIFT GAS

90% Ne - 10%CO2

400 V / cm

E

E

E

E

5.6 m

88ms

NE / CO2

510 cm

500 000 readout channels fabricated, delivered and tested in Lund, now under comissioning in the ALICE magnet

5 m

alice expectations first runs
Commissioning – starting well before first beams arrive

pre-alignment and pre-calibration with cosmic rays

Commissioning with pp events – two stages.

~20k events – used for alignment

first measurements: multiplicity, charges particle spectra

~70M events – used for calibration (TPC, SDD, TOF)

strangeness production, identified particle spectra, particle correlation

First pp runs

~few 109 events

heavy-flavour production, high pT physics, comparison data

ALICE merit, low PT

Pilot heavy-ion run (2008)

~106 events – cover most of the soft physics programme

Full-scale heavy-ion run (2009)

~107 events – first results on hard processes in heavy-ion collisions

Alice expectations first runs