IES Management College And Research Centre

Image from Google Jackets

INTRODUCTION TO ECONOMETRICS G.S. MADDALA AND KAJAL LAHIRI

By: Contributor(s): Publication details: JOHN WILEY & SONS, INC. 2009 NEW DELHIEdition: 4Description: XX, 634 P. PAPERISBN:
  • 978-81-265-3415-9
Subject(s): DDC classification:
  • 330.015195
Contents:
Table of Contents Foreword xvii Preface to the Fourth Edition xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics? 3 1.1 What is econometrics? 3 1.2 Economic and econometric models 4 1.3 The aims and methodology of econometrics 6 1.4 What constitutes a test of an economic theory? 8 CHAPTER 2 Statistical Background and Matrix Algebra 11 2.1 Introduction 11 2.2 Probability 12 2.3 Random variables and probability distributions 17 2.4 The normal probability distribution and related distributions 18 2.5 Classical statistical inference 21 2.6 Properties of estimators 22 2.7 Sampling distributions for samples from a normal population 26 2.8 Interval estimation 26 2.9 Testing of hypotheses 28 2.10 Relationship between confidence interval procedures and tests of hypotheses 31 2.11 Combining independent tests 32 CHAPTER 3 Simple Regression 59 3.1 Introduction 59 3.2 Specification of the relationships 61 3.3 The method of moments 65 3.4 The method of least squares 68 3.5 Statistical inference in the linear regression model 76 3.6 Analysis of variance for the simple regression model 83 3.7 Prediction with the simple regression model 85 3.8 Outliers 88 3.9 Alternative functional forms for regression equations 95 *3.10 Inverse prediction in the least squares regression model1 99 *3.11 Stochastic regressors 102 *3.12 The regression fallacy 102 CHAPTER 4 Multiple Regression 127 4.1 Introduction 127 4.2 A model with two explanatory variables 129 4.3 Statistical inference in the multiple regression model 134 4.4 Interpretation of the regression coefficients 143 4.5 Partial correlations and multiple correlation 146 4.6 Relationships among simple, partial, and multiple correlation coefficients 147 4.7 Prediction in the multiple regression model 153 4.8 Analysis of variance and tests of hypotheses 155 4.9 Omission of relevant variables and inclusion of irrelevant variables 160 4.10 Degrees of freedom and R2 165 4.11 Tests for stability 169 4.12 The LR, W, and LM tests 176 Part II Violation of the Assumptions of the Basic Regression Model 209 CHAPTER 5 Heteroskedasticity 211 5.1 Introduction 211 5.2 Detection of heteroskedasticity 214 5.3 Consequences of heteroskedasticity 219 5.4 Solutions to the heteroskedasticity problem 221 5.5 Heteroskedasticity and the use of deflators 224 5.6 Testing the linear versus log-linear functional form 228 CHAPTER 6 Autocorrelation 239 6.1 Introduction 239 6.2 The Durbin–Watson test 240 6.3 Estimation in levels versus first differences 242 6.4 Estimation procedures with autocorrelated errors 246 6.5 Effect of AR(1) errors on OLS estimates 250 6.6 Some further comments on the DW test 254 6.7 Tests for serial correlation in models with lagged dependent variables 257 6.8 A general test for higher-order serial correlation: The LM test 259 6.9 Strategies when the DW test statistic is significant 261 *6.10 Trends and random walks 266 *6.11 ARCH models and serial correlation 271 6.12 Some comments on the DW test and Durbin’s h-test and t-test 272 CHAPTER 7 Multicollinearity 279 7.1 Introduction 279 7.2 Some illustrative examples 280 7.3 Some measures of multicollinearity 283 7.4 Problems with measuring multicollinearity 286 7.5 Solutions to the multicollinearity problem: Ridge regression 290 7.6 Principal component regression 292 7.7 Dropping variables 297 7.8 Miscellaneous other solutions 300 CHAPTER 8 Dummy Variables and Truncated Variables 313 8.1 Introduction 313 8.2 Dummy variables for changes in the intercept term 314 8.3 Dummy variables for changes in slope coefficients 319 8.4 Dummy variables for cross-equation constraints 322 8.5 Dummy variables for testing stability of regression coefficients 324 8.6 Dummy variables under heteroskedasticity and autocorrelation 327 8.7 Dummy dependent variables 329 8.8 The linear probability model and the linear discriminant function 329 8.9 The probit and logit models 333 8.10 Truncated variables: The tobit model 343 CHAPTER 9 Simultaneous Equation Models 355 9.1 Introduction 355 9.2 Endogenous and exogenous variables 357 9.3 The identification problem: Identification through reduced form 357 9.4 Necessary and sufficient conditions for identification 362 9.5 Methods of estimation: The instrumental variable method 365 9.6 Methods of estimation: The two-stage least squares method 371 9.7 The question of normalization 378 *9.8 The limited-information maximum likelihood method 379 *9.9 On the use of OLS in the estimation of simultaneous equation models 380 *9.10 Exogeneity and causality 386 9.11 Some problems with instrumental variable methods 392 CHAPTER 10 Diagnostic Checking, Model Selection, and Specification Testing 401 10.1 Introduction 401 10.2 Diagnostic tests based on least squares residuals 402 10.3 Problems with least squares residuals 404 10.4 Some other types of residual 405 10.5 DFFITS and bounded influence estimation 411 10.6 Model selection 414 10.7 Selection of regressors 419 10.8 Implied F-ratios for the various criteria 423 10.9 Cross-validation 427 10.10 Hausman’s specification error test 428 10.11 The Plosser–Schwert–White differencing test 435 10.12 Tests for nonnested hypotheses 436 10.13 Nonnormality of errors 440 10.14 Data transformations 441 CHAPTER 11 Errors in Variables 451 11.1 Introduction 451 11.2 The classical solution for a single-equation model with one explanatory variable 452 11.3 The single-equation model with two explanatory variables 455 11.4 Reverse regression 463 11.5 Instrumental variable methods 465 11.6 Proxy variables 468 11.7 Some other problems 471 Part III Special Topics 479 CHAPTER 12 Introduction to Time-Series Analysis 481 12.1 Introduction 481 12.2 Two methods of time-series analysis: Frequency domain and time domain 482 12.3 Stationary and nonstationary time series 482 12.4 Some useful models for time series 485 12.5 Estimation of AR, MA, and ARMA models 492 12.6 The Box–Jenkins approach 496 12.7 R2 measures in time-series models 503 CHAPTER 13 Models of Expectations and Distributed Lags 509 13.1 Models of expectations 509 13.2 Naive models of expectations 510 13.3 The adaptive expectations model 512 13.4 Estimation with the adaptive expectations model 514 13.5 Two illustrative examples 516 13.6 Expectational variables and adjustment lags 520 13.7 Partial adjustment with adaptive expectations 524 13.8 Alternative distributed lag models: Polynomial lags 526 13.9 Rational lags 533 13.10 Rational expectations 534 13.11 Tests for rationality 536 13.12 Estimation of a demand and supply model under rational expectations 538 13.13 The serial correlation problem in rational expectations models 544 CHAPTER 14 Vector Autoregressions, Unit Roots, and Cointegration 551 14.1 Introduction 551 14.2 Vector autoregressions 551 14.3 Problems with VAR models in practice 553 14.4 Unit roots 554 14.5 Unit root tests 555 14.6 Cointegration 563 14.7 The cointegrating regression 564 14.8 Vector autoregressions and cointegration 567 14.9 Cointegration and error correction models 571 14.10 Tests for cointegration 571 14.11 Cointegration and testing of the REH and MEH 572 14.12 A summary assessment of cointegration 574 CHAPTER 15 Panel Data Analysis 583 15.1 Introduction 583 15.2 The LSDV or fixed effects model 584 15.3 The random effects model 586 15.4 Fixed effects versus random effects 589 15.5 Dynamic panel data models 591 15.6 Panel data models with correlated effects and simultaneity 593 15.7 Errors in variables in panel data 595 15.8 The SUR model 597 15.9 The random coefficient model 597 CHAPTER 16 Small-Sample Inference: Resampling Methods 601 16.1 Introduction 601 16.2 Monte Carlo methods 602 16.3 Resampling methods: Jackknife and bootstrap 603 16.4 Bootstrap confidence intervals 605 16.5 Hypothesis testing with the bootstrap 606 16.6 Bootstrapping residuals versus bootstrapping the data 607 16.7 Non-IID errors and nonstationary models 607 Appendix 611 Index 621 See Les
Summary: Maintaining G.S. Maddala’s brilliant expository style of cutting through the technical superstructure to reveal only essential details, while retaining the nerve centre of the subject matter, Professor Kajal Lahiri has brought forward this new edition of one of the most important textbooks in its field. The new edition continues to provide a large number of worked examples, and some shorter data sets. Further data sets and additional supplementary material to assist both the student and lecturer are available on the companion website www.wileyeurope.com/college/maddala
Tags from this library: No tags from this library for this title. Log in to add tags.
Star ratings
    Average rating: 0.0 (0 votes)
Holdings
Item type Current library Collection Call number Status Date due Barcode Item holds
Book Book Main Library ECONOMETRI 330.015195/ MAD/LAH/22806 (Browse shelf(Opens below)) In transit from Main Library to Library Annexe -2 (6th Floor) since 27/02/2024 11122806
Total holds: 0

Table of Contents
Foreword xvii

Preface to the Fourth Edition xix

Part I Introduction and the Linear Regression Model 1

CHAPTER 1 What is Econometrics? 3

1.1 What is econometrics? 3

1.2 Economic and econometric models 4

1.3 The aims and methodology of econometrics 6

1.4 What constitutes a test of an economic theory? 8

CHAPTER 2 Statistical Background and Matrix Algebra 11

2.1 Introduction 11

2.2 Probability 12

2.3 Random variables and probability distributions 17

2.4 The normal probability distribution and related distributions 18

2.5 Classical statistical inference 21

2.6 Properties of estimators 22

2.7 Sampling distributions for samples from a normal population 26

2.8 Interval estimation 26

2.9 Testing of hypotheses 28

2.10 Relationship between confidence interval procedures and tests of hypotheses 31

2.11 Combining independent tests 32

CHAPTER 3 Simple Regression 59

3.1 Introduction 59

3.2 Specification of the relationships 61

3.3 The method of moments 65

3.4 The method of least squares 68

3.5 Statistical inference in the linear regression model 76

3.6 Analysis of variance for the simple regression model 83

3.7 Prediction with the simple regression model 85

3.8 Outliers 88

3.9 Alternative functional forms for regression equations 95

*3.10 Inverse prediction in the least squares regression model1 99

*3.11 Stochastic regressors 102

*3.12 The regression fallacy 102

CHAPTER 4 Multiple Regression 127

4.1 Introduction 127

4.2 A model with two explanatory variables 129

4.3 Statistical inference in the multiple regression model 134

4.4 Interpretation of the regression coefficients 143

4.5 Partial correlations and multiple correlation 146

4.6 Relationships among simple, partial, and multiple correlation coefficients 147

4.7 Prediction in the multiple regression model 153

4.8 Analysis of variance and tests of hypotheses 155

4.9 Omission of relevant variables and inclusion of irrelevant variables 160

4.10 Degrees of freedom and R2 165

4.11 Tests for stability 169

4.12 The LR, W, and LM tests 176

Part II Violation of the Assumptions of the Basic Regression Model 209

CHAPTER 5 Heteroskedasticity 211

5.1 Introduction 211

5.2 Detection of heteroskedasticity 214

5.3 Consequences of heteroskedasticity 219

5.4 Solutions to the heteroskedasticity problem 221

5.5 Heteroskedasticity and the use of deflators 224

5.6 Testing the linear versus log-linear functional form 228

CHAPTER 6 Autocorrelation 239

6.1 Introduction 239

6.2 The Durbin–Watson test 240

6.3 Estimation in levels versus first differences 242

6.4 Estimation procedures with autocorrelated errors 246

6.5 Effect of AR(1) errors on OLS estimates 250

6.6 Some further comments on the DW test 254

6.7 Tests for serial correlation in models with lagged dependent variables 257

6.8 A general test for higher-order serial correlation: The LM test 259

6.9 Strategies when the DW test statistic is significant 261

*6.10 Trends and random walks 266

*6.11 ARCH models and serial correlation 271

6.12 Some comments on the DW test and Durbin’s h-test and t-test 272

CHAPTER 7 Multicollinearity 279

7.1 Introduction 279

7.2 Some illustrative examples 280

7.3 Some measures of multicollinearity 283

7.4 Problems with measuring multicollinearity 286

7.5 Solutions to the multicollinearity problem: Ridge regression 290

7.6 Principal component regression 292

7.7 Dropping variables 297

7.8 Miscellaneous other solutions 300

CHAPTER 8 Dummy Variables and Truncated Variables 313

8.1 Introduction 313

8.2 Dummy variables for changes in the intercept term 314

8.3 Dummy variables for changes in slope coefficients 319

8.4 Dummy variables for cross-equation constraints 322

8.5 Dummy variables for testing stability of regression coefficients 324

8.6 Dummy variables under heteroskedasticity and autocorrelation 327

8.7 Dummy dependent variables 329

8.8 The linear probability model and the linear discriminant function 329

8.9 The probit and logit models 333

8.10 Truncated variables: The tobit model 343

CHAPTER 9 Simultaneous Equation Models 355

9.1 Introduction 355

9.2 Endogenous and exogenous variables 357

9.3 The identification problem: Identification through reduced form 357

9.4 Necessary and sufficient conditions for identification 362

9.5 Methods of estimation: The instrumental variable method 365

9.6 Methods of estimation: The two-stage least squares method 371

9.7 The question of normalization 378

*9.8 The limited-information maximum likelihood method 379

*9.9 On the use of OLS in the estimation of simultaneous equation models 380

*9.10 Exogeneity and causality 386

9.11 Some problems with instrumental variable methods 392

CHAPTER 10 Diagnostic Checking, Model Selection, and Specification Testing 401

10.1 Introduction 401

10.2 Diagnostic tests based on least squares residuals 402

10.3 Problems with least squares residuals 404

10.4 Some other types of residual 405

10.5 DFFITS and bounded influence estimation 411

10.6 Model selection 414

10.7 Selection of regressors 419

10.8 Implied F-ratios for the various criteria 423

10.9 Cross-validation 427

10.10 Hausman’s specification error test 428

10.11 The Plosser–Schwert–White differencing test 435

10.12 Tests for nonnested hypotheses 436

10.13 Nonnormality of errors 440

10.14 Data transformations 441

CHAPTER 11 Errors in Variables 451

11.1 Introduction 451

11.2 The classical solution for a single-equation model with one explanatory variable 452

11.3 The single-equation model with two explanatory variables 455

11.4 Reverse regression 463

11.5 Instrumental variable methods 465

11.6 Proxy variables 468

11.7 Some other problems 471

Part III Special Topics 479

CHAPTER 12 Introduction to Time-Series Analysis 481

12.1 Introduction 481

12.2 Two methods of time-series analysis: Frequency domain and time domain 482

12.3 Stationary and nonstationary time series 482

12.4 Some useful models for time series 485

12.5 Estimation of AR, MA, and ARMA models 492

12.6 The Box–Jenkins approach 496

12.7 R2 measures in time-series models 503

CHAPTER 13 Models of Expectations and Distributed Lags 509

13.1 Models of expectations 509

13.2 Naive models of expectations 510

13.3 The adaptive expectations model 512

13.4 Estimation with the adaptive expectations model 514

13.5 Two illustrative examples 516

13.6 Expectational variables and adjustment lags 520

13.7 Partial adjustment with adaptive expectations 524

13.8 Alternative distributed lag models: Polynomial lags 526

13.9 Rational lags 533

13.10 Rational expectations 534

13.11 Tests for rationality 536

13.12 Estimation of a demand and supply model under rational expectations 538

13.13 The serial correlation problem in rational expectations models 544

CHAPTER 14 Vector Autoregressions, Unit Roots, and Cointegration 551

14.1 Introduction 551

14.2 Vector autoregressions 551

14.3 Problems with VAR models in practice 553

14.4 Unit roots 554

14.5 Unit root tests 555

14.6 Cointegration 563

14.7 The cointegrating regression 564

14.8 Vector autoregressions and cointegration 567

14.9 Cointegration and error correction models 571

14.10 Tests for cointegration 571

14.11 Cointegration and testing of the REH and MEH 572

14.12 A summary assessment of cointegration 574

CHAPTER 15 Panel Data Analysis 583

15.1 Introduction 583

15.2 The LSDV or fixed effects model 584

15.3 The random effects model 586

15.4 Fixed effects versus random effects 589

15.5 Dynamic panel data models 591

15.6 Panel data models with correlated effects and simultaneity 593

15.7 Errors in variables in panel data 595

15.8 The SUR model 597

15.9 The random coefficient model 597

CHAPTER 16 Small-Sample Inference: Resampling Methods 601

16.1 Introduction 601

16.2 Monte Carlo methods 602

16.3 Resampling methods: Jackknife and bootstrap 603

16.4 Bootstrap confidence intervals 605

16.5 Hypothesis testing with the bootstrap 606

16.6 Bootstrapping residuals versus bootstrapping the data 607

16.7 Non-IID errors and nonstationary models 607

Appendix 611

Index 621
See Les

Maintaining G.S. Maddala’s brilliant expository style of cutting through the technical superstructure to reveal only essential details, while retaining the nerve centre of the subject matter, Professor Kajal Lahiri has brought forward this new edition of one of the most important textbooks in its field.

The new edition continues to provide a large number of worked examples, and some shorter data sets. Further data sets and additional supplementary material to assist both the student and lecturer are available on the companion website www.wileyeurope.com/college/maddala

There are no comments on this title.

to post a comment.

Circulation Timings: Monday to Saturday: 8:30 AM to 9:30 PM | Sundays/Bank Holiday during Examination Period: 10:00 AM to 6:00 PM