牙周康又叫什么名字| 动脉硬化吃什么药最好| 组织液是什么| 男人不尊重你说明什么| 脸上长白斑是什么原因| 籽字五行属什么| 舌头无苔是什么原因| 什么叫两会| 钠高是什么原因| 梦见殡仪馆是什么意思| 潮吹是什么意思| 男人都是大猪蹄子是什么意思| 诸葛亮儿子叫什么| 吴亦凡为什么叫牛| 基因是什么意思| 血稠是什么原因引起的| 高湛为什么帮梅长苏| 健脾祛湿吃什么药| 打狂犬疫苗不能吃什么| 始祖是什么意思| 中耳炎有什么症状| 血热吃什么药效果好| 孩子长个子吃什么有利于长高| 兄弟左右来是什么生肖| 唇周围长痘痘是什么原因| 梗阻性黄疸是什么病| 躯体是什么意思| 无垠是什么意思| 心电图j点抬高什么意思| 嗳气吃什么药| 晚上总是睡不着觉是什么原因| 心率过缓吃什么药| 御是什么意思| dcr是什么| 甲醇是什么| 什么也什么造句| 木加一笔有什么字| 牙齿有黑线是什么原因| 双职工是什么意思| 来减一笔是什么字| 头部MRI检查是什么意思| 8023是什么意思| 月是什么意思| 酸奶用什么菌发酵| 慢性前列腺炎有什么症状| 克罗心是什么档次| 相声海清是什么意思| 善对什么| 男士去皱纹用什么好| 西楚霸王是什么生肖| 筋道是什么意思| 日加一笔变成什么字| 梦见大蟒蛇是什么征兆| 亚麻是什么| 西西里的美丽传说讲的什么| 达克宁栓治疗什么妇科病| 肾在什么位置图片| 心肌缺血是什么原因| 中出什么意思| 臭鳜鱼是什么鱼| 玻璃体切除后对眼睛有什么影响| 11.11什么星座| 种植什么药材最赚钱| 茯苓和茯神有什么区别| 紧迫感是什么意思| 吃什么呢| 异地办理护照需要什么材料| 乌鸡汤放什么补气补血| 为什么会气虚| 麦粒肿用什么药| bjd是什么| 12月28是什么星座| 肝硬化什么症状| 气炎念什么| 为什么会厌学| 血常规五项能检查出什么病| 什么是一二三级医院| 羟苯乙酯是什么| 桥本甲状腺炎有什么症状表现| 清高是什么意思| 火鸡面为什么叫火鸡面| 叶赫那拉氏是什么旗| 稻谷什么时候成熟| 胆红素高说明什么| 送男朋友什么礼物合适| 安属于五行属什么| 经常感冒发烧是什么原因| 高血压的人不能吃什么| 怎么吃都不胖是什么原因| 腹部b超可以检查什么| 猪肚搭配什么煲汤最好| 手指甲出现双层是什么原因| 七月八号什么星座| 苯磺酸氨氯地平片什么时候吃| 胎停了有什么明显症状| 天体是什么意思| 臭嗨是什么意思| 纵欲什么意思| 来年是什么意思| 肩周炎口服什么药最好| 鳗鱼吃什么食物| 21度穿什么衣服| 为什么早上起来恶心想吐| 腮腺炎用什么药| 一个口一个犬读什么| 梦见别人家办丧事是什么意思| 什么情况要打破伤风针| 人经常放屁是什么原因| 胎动突然频繁是什么原因| 低回声结节是什么意思| 天狗是什么意思| 耳朵上长痘痘什么原因| leysen是什么牌子| 普洱茶是什么茶| 是故是什么意思| 腿肿吃什么药| 李白为什么叫青莲居士| 白头发有什么方法变黑| 长期便秘喝什么茶好| 现代是什么时候| 什么牌子皮带结实耐用| 三周年祭奠有什么讲究| 尔加玉读什么| 虫草花是什么| 痛风都不能吃什么东西| 身体抱恙是什么意思| 婚检能检查出什么| 拔罐拔出水是什么原因| 井柏然原名叫什么| 釜底抽薪是什么计| 什么龙可以横行霸道| 梦见蛇追我是什么预兆| 水逆退散是什么意思| 长期失眠应该吃什么药| 看痘痘挂什么科| 心什么什么什么| 三个火读什么字| 什么叫阈值| 植发用什么头发| 小t是什么意思| 喝酒后吃什么解酒最快| 沉鱼落雁闭月羞花是什么意思| 吐黑水是什么原因| 鸟字旁的字和什么有关| 为什么会长麦粒肿| 蔗糖脂肪酸酯是什么| 胃胀是什么症状| 葵水是什么意思| 牛剖层皮革是什么意思| 回春是什么意思| 解脲脲原体是什么意思| 口苦吃什么中药| 2013年属什么| 血糖高是什么症状| 木甚念什么| 中老年人补钙吃什么牌子的钙片好| 什么什么不得| 运动后体重增加是什么原因| 撸铁是什么| 劳您费心了什么意思| 男人为什么喜欢吸奶| 吃海带有什么好处| 感冒了喝什么汤好| 白手套是什么意思| 银杏叶片有什么作用| 头发打结是什么原因| 一个虫一个圣念什么| gree是什么牌子| 奎字五行属什么| 吃什么降血压的食物| carrots是什么意思| 原生态是什么意思| 七月十日是什么星座| 病毒发烧吃什么药| 为什么你| 什么植物最老实| 海绵体修复吃什么药| 口蜜什么剑| 幽门螺旋杆菌感染有什么症状| sanag是什么牌子| 牛的本命佛是什么佛| 孩子铅高有什么症状| 多发性结节是什么意思| 什么是尘肺病| 什么叫介入治疗| 天牛喜欢吃什么| 紫花地丁有什么功效| 猫一般吃什么| 牵牛花什么时候开花| 什么叫直男| 镶牙与种牙有什么区别| 白细胞低吃什么食物好| 家里养什么鱼好| 用什么方法治牙疼| 红楼梦是什么朝代| 知我者非你也什么意思| 我的梦想是什么| 精索静脉曲张是什么意思| 小孩什么时候换牙| 顶包是什么意思| pi是什么意思| 势利眼是什么意思| 冷笑是什么意思| 阴虚火旺是什么症状| 姓傅的男孩取什么名字| 梅毒症状男有什么表现| 维生素b6是治什么的| 广州有什么好吃的| 喆读什么| 阴囊瘙痒是什么原因| 指甲看什么科| 腺样体肥大是什么意思| 嘴子是什么意思| 记过属于什么处分| 博爱是什么意思| 鱼泡是鱼的什么器官| 血压高什么原因引起的| 上面日下面立读什么| 早餐做什么| 泪目是什么意思| 手指缝里长水泡还痒是什么原因| 抬头头晕是什么原因| 拔牙为什么要验血| 算什么男人歌词| 孕吐反应什么时候开始| 什么姿势最深| 什么是逆商| 三眼花翎是什么意思| 社保缴费基数是什么意思| 伤口感染吃什么消炎药| 晚上胃疼是什么原因| 喜用神是什么| 名分是什么意思| 12朵玫瑰代表什么意思| 红萝卜什么时候种| 烧心是什么原因引起的| 抽血化验挂什么科| 人流挂什么科| 女人吃什么能增加雌激素| 六味地黄丸有什么用| apc是什么药| 端午节有什么习俗| 老人脚肿是什么原因| 翡翠都有什么颜色| 帕金森吃什么药好得快| 沧海桑田什么意思| 什么是基因检测| 七匹狼男装是什么档次| 人为什么会得阑尾炎| 欣喜若狂的近义词是什么| 海肠是什么东西| 石家庄古代叫什么名字| 财神爷供奉什么供品| 虚岁30岁属什么生肖| 钠氯偏低是什么原因| 梦见自己来月经了什么预兆| cb是什么| 输卵管发炎有什么症状表现| 晚年是什么意思| 一致是什么意思| 慧根是什么意思| 病理检查是什么意思| 什么大什么粗| 摸头是什么意思| 百度

女子用母泰迪“诱拐”8只公泰迪 网友:仙狗跳泰迪女子狗

百度 二是落实专门的活动场所。

In computing, a benchmark is the act of running a computer program, a set of programs, or other operations, in order to assess the relative performance of an object, normally by running a number of standard tests and trials against it.[1]

A graphical demo running as a benchmark of the OGRE engine

The term benchmark is also commonly utilized for the purposes of elaborately designed benchmarking programs themselves.

Benchmarking is usually associated with assessing performance characteristics of computer hardware, for example, the floating point operation performance of a CPU, but there are circumstances when the technique is also applicable to software. Software benchmarks are, for example, run against compilers or database management systems (DBMS).

Benchmarks provide a method of comparing the performance of various subsystems across different chip/system architectures. Benchmarking as a part of continuous integration is called Continuous Benchmarking.[2]

Purpose

edit

As computer architecture advanced, it became more difficult to compare the performance of various computer systems simply by looking at their specifications. Therefore, tests were developed that allowed comparison of different architectures. For example, Pentium 4 processors generally operated at a higher clock frequency than Athlon XP or PowerPC processors, which did not necessarily translate to more computational power; a processor with a slower clock frequency might perform as well as or even better than a processor operating at a higher frequency. See BogoMips and the megahertz myth.

Benchmarks are designed to mimic a particular type of workload on a component or system. Synthetic benchmarks do this by specially created programs that impose the workload on the component. Application benchmarks run real-world programs on the system. While application benchmarks usually give a much better measure of real-world performance on a given system, synthetic benchmarks are useful for testing individual components, like a hard disk or networking device.

Benchmarks are particularly important in CPU design, giving processor architects the ability to measure and make tradeoffs in microarchitectural decisions. For example, if a benchmark extracts the key algorithms of an application, it will contain the performance-sensitive aspects of that application. Running this much smaller snippet on a cycle-accurate simulator can give clues on how to improve performance.

Prior to 2000, computer and microprocessor architects used SPEC to do this, although SPEC's Unix-based benchmarks were quite lengthy and thus unwieldy to use intact.

Computer companies are known to configure their systems to give unrealistically high performance on benchmark tests that are not replicated in real usage. For instance, during the 1980s some compilers could detect a specific mathematical operation used in a well-known floating-point benchmark and replace the operation with a faster mathematically equivalent operation. However, such a transformation was rarely useful outside the benchmark until the mid-1990s, when RISC and VLIW architectures emphasized the importance of compiler technology as it related to performance. Benchmarks are now regularly used by compiler companies to improve not only their own benchmark scores, but real application performance.

CPUs that have many execution units — such as a superscalar CPU, a VLIW CPU, or a reconfigurable computing CPU — typically have slower clock rates than a sequential CPU with one or two execution units when built from transistors that are just as fast. Nevertheless, CPUs with many execution units often complete real-world and benchmark tasks in less time than the supposedly faster high-clock-rate CPU.

Given the large number of benchmarks available, a vendor can usually find at least one benchmark that shows its system will outperform another system; the other systems can be shown to excel with a different benchmark.

Software vendors also use benchmarks in their marketing, such as the "benchmark wars" between rival relational database makers in the 1980s and 1990s. Companies commonly report only those benchmarks (or aspects of benchmarks) that show their products in the best light. They also have been known to mis-represent the significance of benchmarks, again to show their products in the best possible light.[3][4]

Ideally benchmarks should only substitute for real applications if the application is unavailable, or too difficult or costly to port to a specific processor or computer system. If performance is critical, the only benchmark that matters is the target environment's application suite.

Functionality

edit

Features of benchmarking software may include recording/exporting the course of performance to a spreadsheet file, visualization such as drawing line graphs or color-coded tiles, and pausing the process to be able to resume without having to start over. Software can have additional features specific to its purpose, for example, disk benchmarking software may be able to optionally start measuring the disk speed within a specified range of the disk rather than the full disk, measure random access reading speed and latency, have a "quick scan" feature which measures the speed through samples of specified intervals and sizes, and allow specifying a data block size, meaning the number of requested bytes per read request.[5]

Challenges

edit

Benchmarking is not easy and often involves several iterative rounds in order to arrive at predictable, useful conclusions. Interpretation of benchmarking data is also extraordinarily difficult. Here is a partial list of common challenges:

  • Vendors tend to tune their products specifically for industry-standard benchmarks. Norton SysInfo (SI) is particularly easy to tune for, since it mainly biased toward the speed of multiple operations. Use extreme caution in interpreting such results.
  • Some vendors have been accused of "cheating" at benchmarks — designing their systems such that they give much higher benchmark numbers, but are not as effective at the actual likely workload.[6]
  • Many benchmarks focus entirely on the speed of computational performance, neglecting other important features of a computer system, such as:
    • Qualities of service, aside from raw performance. Examples of unmeasured qualities of service include security, availability, reliability, execution integrity, serviceability, scalability (especially the ability to quickly and nondisruptively add or reallocate capacity), etc. There are often real trade-offs between and among these qualities of service, and all are important in business computing. Transaction Processing Performance Council Benchmark specifications partially address these concerns by specifying ACID property tests, database scalability rules, and service level requirements.
    • In general, benchmarks do not measure Total cost of ownership. Transaction Processing Performance Council Benchmark specifications partially address this concern by specifying that a price/performance metric must be reported in addition to a raw performance metric, using a simplified TCO formula. However, the costs are necessarily only partial, and vendors have been known to price specifically (and only) for the benchmark, designing a highly specific "benchmark special" configuration with an artificially low price. Even a tiny deviation from the benchmark package results in a much higher price in real world experience.
    • Facilities burden (space, power, and cooling). When more power is used, a portable system will have a shorter battery life and require recharging more often. A server that consumes more power and/or space may not be able to fit within existing data center resource constraints, including cooling limitations. There are real trade-offs as most semiconductors require more power to switch faster. See also performance per watt.
    • In some embedded systems, where memory is a significant cost, better code density can significantly reduce costs.
  • Vendor benchmarks tend to ignore requirements for development, test, and disaster recovery computing capacity. Vendors only like to report what might be narrowly required for production capacity in order to make their initial acquisition price seem as low as possible.
  • Benchmarks are having trouble adapting to widely distributed servers, particularly those with extra sensitivity to network topologies. The emergence of grid computing, in particular, complicates benchmarking since some workloads are "grid friendly", while others are not.
  • Users can have very different perceptions of performance than benchmarks may suggest. In particular, users appreciate predictability — servers that always meet or exceed service level agreements. Benchmarks tend to emphasize mean scores (IT perspective), rather than maximum worst-case response times (real-time computing perspective), or low standard deviations (user perspective).
  • Many server architectures degrade dramatically at high (near 100%) levels of usage — "fall off a cliff" — and benchmarks should (but often do not) take that factor into account. Vendors, in particular, tend to publish server benchmarks at continuous at about 80% usage — an unrealistic situation — and do not document what happens to the overall system when demand spikes beyond that level.
  • Many benchmarks focus on one application, or even one application tier, to the exclusion of other applications. Most data centers are now implementing virtualization extensively for a variety of reasons, and benchmarking is still catching up to that reality where multiple applications and application tiers are concurrently running on consolidated servers.
  • There are few (if any) high quality benchmarks that help measure the performance of batch computing, especially high volume concurrent batch and online computing. Batch computing tends to be much more focused on the predictability of completing long-running tasks correctly before deadlines, such as end of month or end of fiscal year. Many important core business processes are batch-oriented and probably always will be, such as billing.
  • Benchmarking institutions often disregard or do not follow basic scientific method. This includes, but is not limited to: small sample size, lack of variable control, and the limited repeatability of results.[7]

Benchmarking principles

edit

There are seven vital characteristics for benchmarks.[8] These key properties are:

  1. Relevance: Benchmarks should measure relatively vital features.
  2. Representativeness: Benchmark performance metrics should be broadly accepted by industry and academia.
  3. Equity: All systems should be fairly compared.
  4. Repeatability: Benchmark results can be verified.
  5. Cost-effectiveness: Benchmark tests are economical.
  6. Scalability: Benchmark tests should work across systems possessing a range of resources from low to high.
  7. Transparency: Benchmark metrics should be easy to understand.

Types of benchmark

edit
  1. Real program
  2. Component Benchmark / Microbenchmark
    • core routine consists of a relatively small and specific piece of code.
    • measure performance of a computer's basic components[9]
    • may be used for automatic detection of computer's hardware parameters like number of registers, cache size, memory latency, etc.
  3. Kernel
    • contains key codes
    • normally abstracted from actual program
    • popular kernel: Livermore loop
    • linpack benchmark (contains basic linear algebra subroutine written in FORTRAN language)
    • results are represented in Mflop/s.
  4. Synthetic Benchmark
    • Procedure for programming synthetic benchmark:
      • take statistics of all types of operations from many application programs
      • get proportion of each operation
      • write program based on the proportion above
    • Types of Synthetic Benchmark are:
    • These were the first general purpose industry standard computer benchmarks. They do not necessarily obtain high scores on modern pipelined computers.
  5. I/O benchmarks
  6. Database benchmarks
    • measure the throughput and response times of database management systems (DBMS)
  7. Parallel benchmarks
    • used on machines with multiple cores and/or processors, or systems consisting of multiple machines

Common benchmarks

edit

Industry standard (audited and verifiable)

edit

Open source benchmarks

edit
  • AIM Multiuser Benchmark – composed of a list of tests that could be mixed to create a 'load mix' that would simulate a specific computer function on any UNIX-type OS.
  • Bonnie++ – filesystem and hard drive benchmark
  • BRL-CAD – cross-platform architecture-agnostic benchmark suite based on multithreaded ray tracing performance; baselined against a VAX-11/780; and used since 1984 for evaluating relative CPU performance, compiler differences, optimization levels, coherency, architecture differences, and operating system differences.
  • Collective Knowledge – customizable, cross-platform framework to crowdsource benchmarking and optimization of user workloads (such as deep learning) across hardware provided by volunteers
  • Coremark – Embedded computing benchmark
  • DEISA Benchmark Suite – scientific HPC applications benchmark
  • Dhrystone – integer arithmetic performance, often reported in DMIPS (Dhrystone millions of instructions per second)
  • DiskSpdCommand-line tool for storage benchmarking that generates a variety of requests against computer files, partitions or storage devices
  • Fhourstones – an integer benchmark
  • HINT – designed to measure overall CPU and memory performance
  • Iometer – I/O subsystem measurement and characterization tool for single and clustered systems.
  • IOzone – Filesystem benchmark
  • LINPACK benchmarks – traditionally used to measure FLOPS
  • Livermore loops
  • NAS parallel benchmarks
  • NBench – synthetic benchmark suite measuring performance of integer arithmetic, memory operations, and floating-point arithmetic
  • PAL – a benchmark for realtime physics engines
  • PerfKitBenchmarker – A set of benchmarks to measure and compare cloud offerings.
  • Phoronix Test Suite – open-source cross-platform benchmarking suite for Linux, OpenSolaris, FreeBSD, OSX and Windows. It includes a number of other benchmarks included on this page to simplify execution.
  • POV-Ray – 3D render
  • Tak (function) – a simple benchmark used to test recursion performance
  • TATP Benchmark – Telecommunication Application Transaction Processing Benchmark
  • TPoX – An XML transaction processing benchmark for XML databases
  • VUP (VAX unit of performance) – also called VAX MIPS
  • Whetstone – floating-point arithmetic performance, often reported in millions of Whetstone instructions per second (MWIPS)

Microsoft Windows benchmarks

edit

Unusual benchmark

edit

Others

edit
  • AnTuTu – commonly used on phones and ARM-based devices.
  • Byte Sieve - originally tested language performance, but widely used as a machine benchmark as well.
  • Creative Computing Benchmark – Compares the BASIC programming language on various platforms. Introduced in 1983.
  • Geekbench – A cross-platform benchmark for Windows, Linux, macOS, iOS and Android.
  • iCOMP – the Intel comparative microprocessor performance, published by Intel
  • Khornerstone
  • Novabench - a computer benchmarking utility for Microsoft Windows, macOS, and Linux
  • Performance Rating – modeling scheme used by AMD and Cyrix to reflect the relative performance usually compared to competing products.
  • Rugg/Feldman benchmarks - one of the earliest microcomputer benchmarks, from 1977.
  • SunSpider – a browser speed test
  • UserBenchmark - PC benchmark utility
  • VMmark – a virtualization benchmark suite.

See also

edit

References

edit
  1. ^ Fleming, Philip J.; Wallace, John J. (2025-08-05). "How not to lie with statistics: the correct way to summarize benchmark results". Communications of the ACM. 29 (3): 218–221. doi:10.1145/5666.5673. ISSN 0001-0782. S2CID 1047380.
  2. ^ Grambow, Martin; Lehmann, Fabian; Bermbach, David (2019). "Continuous Benchmarking: Using System Benchmarking in Build Pipelines". 2019 IEEE International Conference on Cloud Engineering (IC2E). pp. 241–246. doi:10.1109/IC2E.2019.00039. ISBN 978-1-7281-0218-4. Retrieved 2025-08-05.
  3. ^ "RDBMS Workshop: Informix" (PDF) (Interview). Interviewed by Luanne Johnson. Computer History Museum. 2025-08-05. Retrieved 2025-08-05.
  4. ^ "RDBMS Workshop: Ingres and Sybase" (PDF) (Interview). Interviewed by Doug Jerger. Computer History Museum. 2025-08-05. Retrieved 2025-08-05.
  5. ^ Software: HDDScan, GNOME Disks
  6. ^ Krazit, Tom (2003). "NVidia's Benchmark Tactics Reassessed". IDG News. Archived from the original on 2025-08-05. Retrieved 2025-08-05.
  7. ^ Castor, Kevin (2006). "Hardware Testing and Benchmarking Methodology". Archived from the original on 2025-08-05. Retrieved 2025-08-05.
  8. ^ Dai, Wei; Berleant, Daniel (December 12–14, 2019). "Benchmarking Contemporary Deep Learning Hardware and Frameworks: a Survey of Qualitative Metrics" (PDF). 2019 IEEE First International Conference on Cognitive Machine Intelligence (CogMI). Los Angeles, CA, USA: IEEE. pp. 148–155. arXiv:1907.03626. doi:10.1109/CogMI48466.2019.00029.
  9. ^ Ehliar, Andreas; Liu, Dake. "Benchmarking network processors" (PDF). {{cite journal}}: Cite journal requires |journal= (help)
  10. ^ Transaction Processing Performance Council (February 1998). "History and Overview of the TPC". TPC. Transaction Processing Performance Council. Retrieved 2025-08-05.

Further reading

edit
edit


散光是什么意思 马蜂蛰了用什么药 奎宁是什么药 下巴底下长痘痘是什么原因 貂蝉姓什么
随喜功德是什么意思 三月初什么星座 怀孕了什么时候做检查 什么什么大什么 土克什么
双肺斑索是什么意思 黑色裤子配什么颜色t恤 为什么我的眼里常含泪水 胎盘什么时候形成 老公护着婆婆说明什么
靶向治疗是什么 鼻塞喉咙痛吃什么药 饭后打嗝是什么原因 幻灭是什么意思 什么流砥柱
1940年中国发生了什么hcv9jop0ns6r.cn 69是什么姿势hcv8jop6ns3r.cn 摆地摊卖什么最赚钱而且很受欢迎hcv9jop2ns6r.cn 脖子淋巴结挂什么科hcv7jop6ns3r.cn 九二年属猴的是什么命gangsutong.com
什么盛开hcv8jop1ns4r.cn 正值当年什么意思hcv7jop5ns6r.cn 神经性头疼吃什么药hcv7jop6ns8r.cn 清朝什么时候灭亡的hcv8jop6ns5r.cn 心脏跳的快吃什么药96micro.com
脚发烫是什么原因hcv8jop5ns6r.cn 酵素是什么hcv8jop9ns1r.cn uu解脲脲原体阳性是什么意思hcv7jop9ns1r.cn ce是什么元素hcv8jop4ns9r.cn 尿酸高中医叫什么病travellingsim.com
盆腔少量积液什么意思hcv7jop5ns0r.cn 牛奶可以做什么美食1949doufunao.com 桃子又什么又什么hcv9jop2ns4r.cn 工科和理科有什么区别hcv7jop6ns2r.cn 怕空调冷风什么原因hcv8jop6ns8r.cn
百度