Biblio

Export 10 results:
Sort by: Author [ Keyword  (Desc)] Title Type Year
Filters: First Letter Of Last Name is H  [Clear All Filters]
A B C D E F G [H] I J K L M N O P Q R S T U V W X Y Z   [Show ALL]
subject motivation
Höst, M, Wohlin C, Thelin T.  2005.  Experimental context classification: incentives and experience of subjects. Proceedings of the 27th international conference on Software engineering. :470–478. Abstract
subject experience
Höst, M, Wohlin C, Thelin T.  2005.  Experimental context classification: incentives and experience of subjects. Proceedings of the 27th international conference on Software engineering. :470–478. Abstract
statistics
Hoefler, T, Belli R.  2015.  Scientific Benchmarking of Parallel Computing Systems: Twelve Ways to Tell the Masses when Reporting Performance Results. Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. :73:1–73:12. Abstract
statistical mistakes
SPEC
Blackburn, SM, Garner R, Hoffmann C, Khang AM, McKinley KS, Bentzur R, Diwan A, Feinberg D, Frampton D, Guyer SZ et al..  2006.  The DaCapo benchmarks: java benchmarking development and analysis. OOPSLA '06: Proceedings of the 21st annual ACM SIGPLAN conference on Object-oriented programming systems, languages, and applications. :169–190. Abstract
software engineering
Hanenberg, S.  2010.  Faith, hope, and love: an essay on software science's neglect of human factors. Proceedings of the ACM international conference on Object oriented programming systems languages and applications. :933–946. Abstract
Scientific evaluation papers
Wieringa, R, Heerkens H, Regnell B.  2009.  How to Write and Read a Scientific Evaluation Paper. Proceedings of the 2009 17th IEEE International Requirements Engineering Conference, RE. :361–364. Abstract
Research reporting
Wieringa, R, Heerkens H, Regnell B.  2009.  How to Write and Read a Scientific Evaluation Paper. Proceedings of the 2009 17th IEEE International Requirements Engineering Conference, RE. :361–364. Abstract
research methods
Hanenberg, S.  2010.  Faith, hope, and love: an essay on software science's neglect of human factors. Proceedings of the ACM international conference on Object oriented programming systems languages and applications. :933–946. Abstract
Research methodology
Wieringa, R, Heerkens H, Regnell B.  2009.  How to Write and Read a Scientific Evaluation Paper. Proceedings of the 2009 17th IEEE International Requirements Engineering Conference, RE. :361–364. Abstract
research guidelines
programming language research
Hanenberg, S.  2010.  Faith, hope, and love: an essay on software science's neglect of human factors. Proceedings of the ACM international conference on Object oriented programming systems languages and applications. :933–946. Abstract
Performance
Mytkowicz, T, Diwan A, Hauswirth M, Sweeney PF.  2009.  Producing wrong data without doing anything obviously wrong!. ASPLOS '09: Proceeding of the 14th international conference on Architectural support for programming languages and operating systems. :265–276. Abstract
Mytkowicz, T, Diwan A, Hauswirth M, Sweeney PF.  2010.  Evaluating the accuracy of Java profilers. PLDI '10: Proceedings of the 2010 ACM SIGPLAN conference on Programming language design and implementation. :187–197. Abstract
parallel computing
Hoefler, T, Belli R.  2015.  Scientific Benchmarking of Parallel Computing Systems: Twelve Ways to Tell the Masses when Reporting Performance Results. Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. :73:1–73:12. Abstract
observation study
methodology
Blackburn, SM, Garner R, Hoffmann C, Khang AM, McKinley KS, Bentzur R, Diwan A, Feinberg D, Frampton D, Guyer SZ et al..  2006.  The DaCapo benchmarks: java benchmarking development and analysis. OOPSLA '06: Proceedings of the 21st annual ACM SIGPLAN conference on Object-oriented programming systems, languages, and applications. :169–190. Abstract
Measurement
Mytkowicz, T, Diwan A, Hauswirth M, Sweeney PF.  2009.  Producing wrong data without doing anything obviously wrong!. ASPLOS '09: Proceeding of the 14th international conference on Architectural support for programming languages and operating systems. :265–276. Abstract
Java
Blackburn, SM, Garner R, Hoffmann C, Khang AM, McKinley KS, Bentzur R, Diwan A, Feinberg D, Frampton D, Guyer SZ et al..  2006.  The DaCapo benchmarks: java benchmarking development and analysis. OOPSLA '06: Proceedings of the 21st annual ACM SIGPLAN conference on Object-oriented programming systems, languages, and applications. :169–190. Abstract
Experimentation
Mytkowicz, T, Diwan A, Hauswirth M, Sweeney PF.  2009.  Producing wrong data without doing anything obviously wrong!. ASPLOS '09: Proceeding of the 14th international conference on Architectural support for programming languages and operating systems. :265–276. Abstract
Höst, M, Wohlin C, Thelin T.  2005.  Experimental context classification: incentives and experience of subjects. Proceedings of the 27th international conference on Software engineering. :470–478. Abstract
Mytkowicz, T, Diwan A, Hauswirth M, Sweeney PF.  2010.  Evaluating the accuracy of Java profilers. PLDI '10: Proceedings of the 2010 ACM SIGPLAN conference on Programming language design and implementation. :187–197. Abstract
experimental infrastructure
Experimental evaluation