论文精读--GPT4

现有的所有模型都无法做到在线学习,能力有限,而让大模型拥有一个tools工具库,则可以使大模型变成一个交互式的工具去协调调用API完成任务,同时GPT4还联网了,可以不断地更新自己的知识库

多模态模型,接受文本、图像的输入

由于GPT4论文展现的技术细节较少,安全性展示较少,所以不做精读,而是读一篇openai官网对GPT4的一份调查报告

GPT-4 (openai.com)

We’ve created GPT-4, the latest milestone in OpenAI’s effort in scaling up deep learning. GPT-4 is a large multimodal model (accepting image and text inputs, emitting text outputs) that, while less capable than humans in many real-world scenarios, exhibits human-level performance on various professional and academic benchmarks. For example, it passes a simulated bar exam with a score around the top 10% of test takers; in contrast, GPT-3.5’s score was around the bottom 10%. We’ve spent 6 months iteratively aligning GPT-4 using lessons from our adversarial testing program as well as ChatGPT, resulting in our best-ever results (though far from perfect) on factuality, steerability, and refusing to go outside of guardrails.

翻译:

我们创建了GPT-4,这是开放AI在深度学习规模化方面的最新里程碑。GPT-4是一个大型多模态模型(接受图像和文本输入,输出文本),虽然在许多现实场景中不如人类那样具有高超的能力,但在各种专业和学术基准测试中表现出人类水平的性能。例如,它在模拟的律师资格考试中的得分约为测试者中前10%;相比之下,GPT-3.5的得分约为测试者中最后10%。我们花费了6个月的时间,通过从我们的对抗测试计划和ChatGPT中得到的经验教训,迭代地调整GPT-4,从而取得了我们有史以来在事实性、可操纵性和拒绝超出范围方面的最佳结果(尽管还远非完美)。

总结:

效果好;从去年8月以来做了很多安全性测试

Over the past two years, we rebuilt our entire deep learning stack and, together with Azure, co-designed a supercomputer from the ground up for our workload. A year ago, we trained GPT-3.5 as a first “test run” of the system. We found and fixed some bugs and improved our theoretical foundations. As a result, our GPT-4 training run was (for us at least!) unprecedentedly stable, becoming our first large model whose training performance we were able to accurately predict ahead of time. As we continue to focus on reliable scaling, we aim to hone our methodology to help us predict and prepare for future capabilities increasingly far in advance—something we view as critical for safety.

翻译:

在过去的两年里,我们重新构建了整个深度学习堆栈,并与Azure合作,从头开始设计了一台超级计算机以适应我们的工作负载。一年前,我们训练了GPT-3.5作为系统的第一个“测试运行”。我们发现并修复了一些错误,并改进了我们的理论基础。结果,我们的GPT-4训练运行(至少对我们来说)是前所未有的稳定,成为我们能够提前准确预测其训练性能的第一个大型模型。随着我们继续专注于可靠的扩展,我们的目标是完善我们的方法论,以帮助我们预测和准备未来能力,这对安全至关重要。

总结:

GPT4可以准确预测本次训练的预期结果,不必等到模型完全训练完成才能知道这组参数有没有用,想法有没有work

可以精准预测在小模型上做的消融实验放到大模型上的结果,而不会受到涌现的影响

Capabilities

In a casual conversation, the distinction between GPT-3.5 and GPT-4 can be subtle. The difference comes out when the complexity of the task reaches a sufficient threshold—GPT-4 is more reliable, creative, and able to handle much more nuanced instructions than GPT-3.5.

To understand the difference between the two models, we tested on a variety of benchmarks, including simulating exams that were originally designed for humans. We proceeded by using the most recent publicly-available tests (in the case of the Olympiads and AP free response questions) or by purchasing 2022–2023 editions of practice exams. We did no specific training for these exams. A minority of the problems in the exams were seen by the model during training, but we believe the results to be representative—see our technical report for details.

翻译:

在非正式的对话中,GPT-3.5和GPT-4之间的区别可能是微妙的。当任务的复杂性达到足够的阈值时,这种区别就会显现出来——相比于GPT-3.5,GPT-4更可靠、更具创造力,能够处理比较微妙的指令。

为了理解这两个模型之间的区别,我们在各种基准测试中进行了测试,包括模拟最初设计给人类的考试。我们首先使用最近公开可用的测试(在奥林匹克和AP自由回答问题的情况下),或购买了2022-2023年版的练习考试。我们并未为这些考试进行特定的训练。在考试中,只有少数问题是模型在训练期间见过的,但我们相信结果是具有代表性的——详细信息请参阅我们的技术报告。

总结:

找了些bechmark来比较3.5和4的能力

深绿色是加上了图片输入,效果提升了

GPT系列在数学方面还是不行

We also evaluated GPT-4 on traditional benchmarks designed for machine learning models. GPT-4 considerably outperforms existing large language models, alongside most state-of-the-art (SOTA) models which may include benchmark-specific crafting or additional training protocols:

翻译:

我们还在为机器学习模型设计的传统基准测试上评估了GPT-4。与现有的大型语言模型以及大多数最先进的模型相比,GPT-4表现出了显著的优势,这些模型可能包括特定于基准测试的设计或额外的训练协议:

总结:

刷一刷之前文本领域的bechmark,全面超过LM的SOTA

 Many existing ML benchmarks are written in English. To get an initial sense of capability in other languages, we translated the MMLU benchmark—a suite of 14,000 multiple-choice problems spanning 57 subjects—into a variety of languages using Azure Translate (see Appendix). In the 24 of 26 languages tested, GPT-4 outperforms the English-language performance of GPT-3.5 and other LLMs (Chinchilla, PaLM), including for low-resource languages such as Latvian, Welsh, and Swahili:

翻译:

许多现有的机器学习基准测试都是用英文编写的。为了初步了解在其他语言中的能力,我们使用Azure翻译将MMLU基准测试(包含14,000个跨越57个学科的多选题)翻译成了多种语言(详见附录)。在测试的26种语言中的24种语言中,GPT-4在性能上优于GPT-3.5和其他大型语言模型(如Chinchilla、PaLM)的英文表现,包括对拉脱维亚语、威尔士语和斯瓦希里语等资源匮乏的语言:

总结:

 Visual inputs

GPT-4 can accept a prompt of text and images, which—parallel to the text-only setting—lets the user specify any vision or language task. Specifically, it generates text outputs (natural language, code, etc.) given inputs consisting of interspersed text and images. Over a range of domains—including documents with text and photographs, diagrams, or screenshots—GPT-4 exhibits similar capabilities as it does on text-only inputs. Furthermore, it can be augmented with test-time techniques that were developed for text-only language models, including few-shot and chain-of-thought prompting. Image inputs are still a research preview and not publicly available.

翻译:

GPT-4可以接受包含文本和图像的提示,这与仅文本设置相对应,使用户能够指定任何视觉或语言任务。具体而言,它能够生成文本输出(自然语言、代码等),给定由交错的文本和图像组成的输入。在包括文本和照片、图表或屏幕截图在内的一系列领域中,GPT-4表现出与仅文本输入相似的能力。此外,它还可以利用针对仅文本语言模型开发的测试时间技术进行增强,包括少样本学习和思维链提示。图像输入仍处于研究预览阶段,尚未公开提供。

总结:

让GPT4说出图片搞笑的点

先对图片做OCR,然后讲法语翻译成英文,然后解题

让GPT读论文,并总结

We preview GPT-4’s performance by evaluating it on a narrow suite of standard academic vision benchmarks. However, these numbers do not fully represent the extent of its capabilities as we are constantly discovering new and exciting tasks that the model is able to tackle. We plan to release further analyses and evaluation numbers as well as thorough investigation of the effect of test-time techniques soon.

翻译:

在评估GPT-4时,我们使用了一系列标准学术视觉基准测试,但这些数据并不能完全代表其能力的广泛程度,因为我们不断发现模型能够处理新的、令人兴奋的任务。我们计划很快发布更多的分析和评估数据,以及对测试时间技术效果的彻底调查。

总结:

在视觉多模态领域比较

Steerability

We’ve been working on each aspect of the plan outlined in our post about defining the behavior of AIs, including steerability. Rather than the classic ChatGPT personality with a fixed verbosity, tone, and style, developers (and soon ChatGPT users) can now prescribe their AI’s style and task by describing those directions in the “system” message. System messages allow API users to significantly customize their users’ experience within bounds. We will keep making improvements here (and particularly know that system messages are the easiest way to “jailbreak” the current model, i.e., the adherence to the bounds is not perfect), but we encourage you to try it out and let us know what you think.

翻译:

我们一直在执行我们在关于定义AI行为的帖子中概述的计划的每个方面,包括可操控性。与固定的冗长度、语气和风格的经典ChatGPT个性不同,开发人员(以及很快会是ChatGPT用户)现在可以通过在“系统”消息中描述这些方向来指定他们的AI的风格和任务。系统消息允许API用户在一定范围内显着自定义他们用户的体验。我们将继续在这方面进行改进(特别是要知道系统消息是“越狱”当前模型的最简单方式,即,对边界的遵守并不完美),但我们鼓励您尝试一下,并告诉我们您的想法。

总结:

System Message可以定义AI用什么样的语气和你说话

Limitations

Despite its capabilities, GPT-4 has similar limitations as earlier GPT models. Most importantly, it still is not fully reliable (it “hallucinates” facts and makes reasoning errors). Great care should be taken when using language model outputs, particularly in high-stakes contexts, with the exact protocol (such as human review, grounding with additional context, or avoiding high-stakes uses altogether) matching the needs of a specific use-case.

While still a real issue, GPT-4 significantly reduces hallucinations relative to previous models (which have themselves been improving with each iteration). GPT-4 scores 40% higher than our latest GPT-3.5 on our internal adversarial factuality evaluations:

翻译:

尽管具备了强大的功能,但GPT-4仍然存在着与之前的GPT模型类似的局限性。最重要的是,它仍然不能完全可靠(会“幻想”事实并产生推理错误)。在使用语言模型输出时应该格外小心,特别是在高风险的情境中,确切的协议(如人工审查、使用额外的上下文来进行确认,或者完全避免高风险的使用)应该根据特定用例的需求来确定。

尽管这仍然是一个真正的问题,但相对于以往的模型(它们自身在每次迭代中都有所改进),GPT-4显著减少了幻觉现象。在我们内部对抗性事实性评估中,GPT-4的得分比我们最新的GPT-3.5高出40%:

总结:

幻觉

The model can have various biases in its outputs—we have made progress on these but there’s still more to do. Per our recent blog post, we aim to make AI systems we build have reasonable default behaviors that reflect a wide swathe of users’ values, allow those systems to be customized within broad bounds, and get public input on what those bounds should be.

GPT-4 generally lacks knowledge of events that have occurred after the vast majority of its data cuts off (September 2021), and does not learn from its experience. It can sometimes make simple reasoning errors which do not seem to comport with competence across so many domains, or be overly gullible in accepting obvious false statements from a user. And sometimes it can fail at hard problems the same way humans do, such as introducing security vulnerabilities into code it produces.

GPT-4 can also be confidently wrong in its predictions, not taking care to double-check work when it’s likely to make a mistake. Interestingly, the base pre-trained model is highly calibrated (its predicted confidence in an answer generally matches the probability of being correct). However, through our current post-training process, the calibration is reduced.

翻译:

模型在输出中可能存在各种偏见——我们在这方面已经取得了进展,但还有更多工作要做。根据我们最近的博客文章,我们的目标是使我们构建的AI系统具有合理的默认行为,反映了广泛用户价值观的范围,允许在广泛范围内定制这些系统,并征求公众对这些边界应该是什么的意见。

GPT-4通常缺乏对其数据大部分截止时间(2021年9月)之后发生的事件的了解,并且不从其经验中学习。它有时会产生简单的推理错误,这与在如此多的领域中表现出的能力不一致,或者在接受用户明显错误的陈述时过于轻信。有时它会以与人类相同的方式在难题上失败,比如在生成的代码中引入安全漏洞。

GPT-4在预测中也可能会自信满满地犯错,在可能犯错时不进行仔细检查。有趣的是,基础的预训练模型是高度校准的(它对答案的预测置信度通常与正确的概率相匹配)。然而,通过我们目前的后训练过程,校准程度会降低。

总结:

偏见

训练数据有时限性

对人服从,过于轻信

模型对其预测的信心与正确概率密切匹配。虚线对角线代表完美的校准。

经过RLHF后反而不行了,更有主观性了

Risks & mitigations

We’ve been iterating on GPT-4 to make it safer and more aligned from the beginning of training, with efforts including selection and filtering of the pretraining data, evaluations and expert engagement, model safety improvements, and monitoring and enforcement.

GPT-4 poses similar risks as previous models, such as generating harmful advice, buggy code, or inaccurate information. However, the additional capabilities of GPT-4 lead to new risk surfaces. To understand the extent of these risks, we engaged over 50 experts from domains such as AI alignment risks, cybersecurity, biorisk, trust and safety, and international security to adversarially test the model. Their findings specifically enabled us to test model behavior in high-risk areas which require expertise to evaluate. Feedback and data from these experts fed into our mitigations and improvements for the model; for example, we’ve collected additional data to improve GPT-4’s ability to refuse requests on how to synthesize dangerous chemicals.

GPT-4 incorporates an additional safety reward signal during RLHF training to reduce harmful outputs (as defined by our usage guidelines) by training the model to refuse requests for such content. The reward is provided by a GPT-4 zero-shot classifier judging safety boundaries and completion style on safety-related prompts. To prevent the model from refusing valid requests, we collect a diverse dataset from various sources (e.g., labeled production data, human red-teaming, model-generated prompts) and apply the safety reward signal (with a positive or negative value) on both allowed and disallowed categories. 

Our mitigations have significantly improved many of GPT-4’s safety properties compared to GPT-3.5. We’ve decreased the model’s tendency to respond to requests for disallowed content by 82% compared to GPT-3.5, and GPT-4 responds to sensitive requests (e.g., medical advice and self-harm) in accordance with our policies 29% more often.

翻译:

我们一直在对GPT-4进行迭代,从训练开始就使其更安全和更对齐,包括对预训练数据的选择和过滤、评估和专家参与、模型安全改进以及监控和执行等方面的努力。

GPT-4存在与之前模型类似的风险,例如生成有害建议、错误的代码或不准确的信息。然而,GPT-4的额外功能导致了新的风险面。为了了解这些风险的程度,我们与来自AI对齐风险、网络安全、生物风险、信任与安全以及国际安全等领域的50多位专家进行了对抗性测试。他们的发现特别让我们能够测试模型在需要专业知识评估的高风险领域中的行为。这些专家的反馈和数据为我们的模型缓解和改进提供了支持;例如,我们收集了额外的数据以提高GPT-4拒绝合成危险化学物质的能力。

在RLHF训练期间,GPT-4集成了一个额外的安全奖励信号,以减少有害输出(根据我们的使用准则定义),通过训练模型拒绝对此类内容的请求。奖励由一个GPT-4零样分类器提供,用于判断安全边界和安全相关提示的完成样式。为了防止模型拒绝有效请求,我们从各种来源收集了多样化的数据集(例如,标记的生产数据、人类红队测试、模型生成的提示)并在允许和不允许的类别上应用安全奖励信号(具有正值或负值)。

我们的缓解措施相比于GPT-3.5显著改善了GPT-4的许多安全属性。与GPT-3.5相比,我们将模型对不允许内容的请求的反应倾向减少了82%,而GPT-4对敏感请求(例如医疗建议和自伤行为)的回应频率根据我们的政策提高了29%。

总结:

(1)人力收集数据提高安全性

(2)利用自己,做了个reward signal,是从预训练好的gpt4中拿出来对prompt进行sensitive检测

Overall, our model-level interventions increase the difficulty of eliciting bad behavior but doing so is still possible. Additionally, there still exist “jailbreaks” to generate content which violate our usage guidelines. As the “risk per token” of AI systems increases, it will become critical to achieve extremely high degrees of reliability in these interventions; for now it’s important to complement these limitations with deployment-time safety techniques like monitoring for abuse.

GPT-4 and successor models have the potential to significantly influence society in both beneficial and harmful ways. We are collaborating with external researchers to improve how we understand and assess potential impacts, as well as to build evaluations for dangerous capabilities that may emerge in future systems. We will soon share more of our thinking on the potential social and economic impacts of GPT-4 and other AI systems.

翻译:

总体而言,我们在模型层面的干预措施增加了引发不良行为的难度,但这仍然是可能的。此外,仍然存在“越狱”的方法来生成违反我们使用准则的内容。随着AI系统的“风险每个令牌”的增加,在这些干预措施中实现极高的可靠性将变得至关重要;目前,将这些限制与部署时的安全技术相结合,例如监控滥用情况,是非常重要的。

GPT-4和后续模型有潜力在社会上产生显着的积极和消极影响。我们正在与外部研究人员合作,改进我们对潜在影响的理解和评估方法,以及为可能在未来系统中出现的危险功能构建评估方法。我们很快将分享更多关于GPT-4和其他AI系统潜在社会和经济影响的思考。

Training process

Like previous GPT models, the GPT-4 base model was trained to predict the next word in a document, and was trained using publicly available data (such as internet data) as well as data we’ve licensed. The data is a web-scale corpus of data including correct and incorrect solutions to math problems, weak and strong reasoning, self-contradictory and consistent statements, and representing a great variety of ideologies and ideas.

翻译:

与先前的GPT模型一样,GPT-4基础模型被训练以预测文档中的下一个词,并且使用了公开可用的数据(如互联网数据)以及我们许可的数据进行训练。这些数据是一个网络规模的数据语料库,包括数学问题的正确和错误解决方案,弱和强的推理,自相矛盾和一致的陈述,代表了各种意识形态和观念。

So when prompted with a question, the base model can respond in a wide variety of ways that might be far from a user’s intent. To align it with the user’s intent within guardrails, we fine-tune the model’s behavior using reinforcement learning with human feedback (RLHF).

翻译:

在面对问题时,基础模型可能会以多种与用户意图相距甚远的方式进行回应。为了使其在用户意图的框架内对齐,我们使用强化学习与人类反馈(RLHF)对模型的行为进行微调。

总结:

基础模型有的时候的回答会和人想要的回答相差很远,所以用RLHF的技术微调了一下

Note that the model’s capabilities seem to come primarily from the pre-training process—RLHF does not improve exam performance (without active effort, it actually degrades it). But steering of the model comes from the post-training process—the base model requires prompt engineering to even know that it should answer the questions.

翻译:

请注意,模型的能力似乎主要来自预训练过程——强化学习与人类反馈并未改善考试表现(没有主动的努力,实际上它会降低考试表现)。但是,对模型的引导来自于后训练过程——基础模型需要及时的工程处理,甚至知道它应该回答这些问题。

总结:

虽然RLHF并没有带来很好的提分,但RLHF还是控制了模型生成人更愿意接受的回答方式

Predictable scaling

A large focus of the GPT-4 project has been building a deep learning stack that scales predictably. The primary reason is that, for very large training runs like GPT-4, it is not feasible to do extensive model-specific tuning. We developed infrastructure and optimization that have very predictable behavior across multiple scales. To verify this scalability, we accurately predicted in advance GPT-4’s final loss on our internal codebase (not part of the training set) by extrapolating from models trained using the same methodology but using 10,000x less compute:

翻译:

GPT-4项目的一个主要关注点是构建一个可预测扩展的深度学习堆栈。主要原因是,对于像GPT-4这样的非常大的训练运行,进行大量的模型特定调整是不可行的。我们开发了基础设施和优化,这些基础设施和优化在多个规模上表现出非常可预测的行为。为了验证这种可扩展性,我们通过从使用相同方法论进行训练但计算量减少了10,000倍的模型进行外推,准确预测了GPT-4在我们的内部代码库(不属于训练集)上的最终损失:

总结:

这么大的模型不可能去做大规模的调参

就算有大量的机器并行去跑,loss也容易跑飞

OpenAI研发了一套infra和优化方法,在多个尺度上实现了训练的稳定性,刚开始训练的时候就已经能预测出最终的loss了

这里可以看出GPT4确实很好地拟合出了loss的曲线

Some capabilities are still hard to predict. For example, the Inverse Scaling Prize was a competition to find a metric that gets worse as model compute increases, and hindsight neglect was one of the winners. Just like with another recent result, GPT-4 reverses the trend:

翻译:

一些能力仍然很难预测。例如,反比例缩放奖是一个竞赛,旨在找到一个随着模型计算增加而变差的指标,而事后忽视就是其中之一的获奖者。就像另一个最近的结果一样,GPT-4扭转了这一趋势:

总结:

这个competition是GPT3出的时候故意找的,用来证明大模型不是所有任务都比小模型要好

牛头不对马嘴的任务中GPT4反而出现了非理性的选择,跳出了逻辑,为结果服务

API

gpt-4 has a context length of 8,192 tokens. We are also providing limited access to our 32,768–context (about 50 pages of text) version, gpt-4-32k, which will also be updated automatically over time (current version gpt-4-32k-0314, also supported until June 14). Pricing is $0.06 per 1K prompt tokens and $0.12 per 1k completion tokens. We are still improving model quality for long context and would love feedback on how it performs for your use-case. We are processing requests for the 8K and 32K engines at different rates based on capacity, so you may receive access to them at different times.

翻译:

GPT-4的上下文长度为8,192个标记。我们还提供对我们的32,768个标记(约50页文本)版本gpt-4-32k的有限访问权限,该版本也将随时间自动更新(当前版本为gpt-4-32k-0314,支持至6月14日)。定价为每1,000个提示标记0.06美元,每1,000个完成标记0.12美元。我们仍在改进长上下文的模型质量,并希望了解它在您的使用案例中的表现。我们根据容量以不同的速度处理对8K和32K引擎的请求,因此您可能会在不同的时间收到对它们的访问权限。

Conclusion

We look forward to GPT-4 becoming a valuable tool in improving people’s lives by powering many applications. There’s still a lot of work to do, and we look forward to improving this model through the collective efforts of the community building on top of, exploring, and contributing to the model.

翻译:

我们期待着GPT-4成为一个有价值的工具,通过为许多应用程序提供支持来改善人们的生活。还有很多工作要做,我们期待着通过社区的集体努力,不断改进这个模型,探索和为模型做出贡献。

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处:/a/510447.html

如若内容造成侵权/违法违规/事实不符,请联系我们进行投诉反馈qq邮箱809451989@qq.com,一经查实,立即删除!

相关文章

面经分享(Flask,轻量级Web框架)

1. Flask的核心特点 a. 轻量级:核心简洁,只提供了基本的功能,其他高级功能可以通过插件或扩展来添加。 b. 灵活性:允许开发者选择适合自己项目的组件和工具,没有强制的项目结构和设计模式。 c. 易于扩展:提…

JVM 内存溢出排查

说明:记录一次JVM内存溢出的排查过程; 场景 项目开发完成后,首次提交到测试环境。测试、产品同事反馈页面先是操作响应慢,抛出超时异常,最后直接无法使用。查看日志后得知是内存溢出。 重启服务后,我对前…

接口测试实战教程(加密解密攻防)

🍅 视频学习:文末有免费的配套视频可观看 🍅 关注公众号【互联网杂货铺】,回复 1 ,免费获取软件测试全套资料,资料在手,涨薪更快 一、对称加密 对称加密算法是共享密钥加密算法,在加…

RuntimeError: Error compiling objects for extension虚拟环境和系统环境——添加、删除、修改环境变量

前言:因为一个报错RuntimeError: Error compiling objects for extension 没有配置cl.exe环境变量,我的应用场景是需要搞定虚拟环境变量配置 RuntimeError: Error compiling objects for extension手把手带你解决(超详细)-CSDN博…

Mybatis 之 useGeneratedKeys

数据库中主键基本都设置为自增,当我们要插入一条数据想要获取这条数据的 Id 时,就可使用 Mybatis 中的 useGeneratedKeys 属性。 背景 这里以 苍穹外卖 中的 新增菜品 功能为例,有 菜品表(dish table)和 口味表(dish_flavor table)&#xf…

HTML标签之有序列表,无序列表,自定义链表

无序列表 <!DOCTYPE html> <html lang"en"> <head><meta charset"UTF-8"><meta http-equiv"X-UA-Compatible" content"IEedge"><meta name"viewport" content"widthdevice-width, i…

从零开始学数据分析之数据分析概述

当今世界对信息技术的依赖程度在不断加深&#xff0c;每天都会有大量的数据产生&#xff0c;我们经常会感到数据越来越多&#xff0c;但是要从中发现有价值的信息却越来越难。 这里所说的信息&#xff0c;可以理解为对数据集处理之后的结果&#xff0c;是从数据集中提炼出的可用…

redis 高可用与 持久化

目录 一 redis 高可用 &#xff08;一&#xff09;redis高可用技术 1&#xff0c;持久化 2&#xff0c;主从复制 3&#xff0c;哨兵 4&#xff0c;Cluster集群 二 redis 持久化 &#xff08;一&#xff09;持久化相关概念介绍 1&#xff0c;持久化的功能 2&…

Pygame基础9-射击

简介 玩家用鼠标控制飞机&#xff08;白色方块&#xff09;移动&#xff0c;按下鼠标后&#xff0c;玩家所在位置出现子弹&#xff0c;子弹匀速向右飞行。 代码 没有什么新的东西&#xff0c;使用两个精灵类表示玩家和子弹。 有一个细节需要注意&#xff0c;当子弹飞出屏幕…

实战 | YOLOv8自定义数据集训练实现手势识别 (标注+训练+预测 保姆级教程--含数据集)

导 读 本文将手把手教你用YoloV8训练自己的数据集并实现手势识别。 安装环境 【1】安装torch, torchvision对应版本,这里先下载好,直接安装 pip install torch-1.13.1+cu116-cp38-cp38-win_amd64.whlpip install torchvision-0.14.1+cu116-cp38-cp38-win_amd64.whl 安装好…

剑指offer打卡 JZ7 重建二叉树

在牛客网刷的&#xff0c;还是跟leetcode一样非acm模式&#xff0c;由于急着暑期实习题量不固定&#xff0c;八股算法轮刷 打卡内容偏个人笔记&#xff0c;本人水平一般(代码随想录稀里糊涂刷了一遍)&#xff0c;从小白开始分析(甚至会分析语法)&#xff0c;尽量一题多解深入探…

VSCode好用插件

由于现在还是使用vue2&#xff0c;所以本文只记录vue2开发中好用的插件。 美化类插件不介绍了&#xff0c;那些貌似对生产力起不到什么大的帮助&#xff0c;纯粹的“唯心主义”罢了&#xff0c;但是如果你有兴趣的话可以查看上一篇博客&#xff1a;VSCode美化 1. vuter 简介&…

LabelConvert: 目标检测和图像分割数据集格式转换工具

LabelConvert LabelConvert是一个目标检测和图像分割的数据集格式转换工具&#xff0c;支持labelme、labelImg与YOLO、VOC和COCO 数据集格式之间的相互转换。 支持的转换格式 安装 pip install label_convert具体使用方法 由于文章篇幅所限&#xff0c;请移步LabelConvert官…

左值与右值,以及c++11的相关特性。

目录 左值 右值 左值引用总结&#xff1a; 右值引用总结&#xff1a; 右值引用使用场景和意义&#xff1a; 1、左值引用的使用场景&#xff1a; 编译器优化1&#xff1a; 2、移动构造与移动赋值&#xff1a; 3、右值引用的使用场景&#xff1a; 编译器优化2&#xff1a…

MATLAB科研绘图与学术图表绘制从入门到精通

&#x1f482; 个人网站:【 摸鱼游戏】【神级代码资源网站】【工具大全】&#x1f91f; 一站式轻松构建小程序、Web网站、移动应用&#xff1a;&#x1f449;注册地址&#x1f91f; 基于Web端打造的&#xff1a;&#x1f449;轻量化工具创作平台&#x1f485; 想寻找共同学习交…

软件资源分享六:EPLAN Electric P8 2024 | Eplan 2024 中文版软件介绍+保姆级安装教程

原文链接&#xff1a;安装激活教程 EPLAN Electric P8 2024 | Eplan 2024 中文版软件介绍安装教程 EPLAN 2024是一款电气设计软件&#xff0c;它可以用于自动化系统的设计、文档编制和维护。EPLAN可以对电气设计的各个方面进行完整的支持&#xff0c;包括电气控制系统、机械设…

新手学python还是c?

考虑到个人情况和职业规划是非常重要的。我这里有一套编程入门教程&#xff0c;不仅包含了详细的视频讲解&#xff0c;项目实战。如果你渴望学习编程&#xff0c;不妨点个关注&#xff0c;给个评论222&#xff0c;私信22&#xff0c;我在后台发给你。 Python作为初学者入门语言…

基于单片机温湿度PM2.5报警设置系统

**单片机设计介绍&#xff0c;基于单片机温湿度PM2.5报警设置系统 文章目录 一 概要二、功能设计设计思路 三、 软件设计原理图 五、 程序六、 文章目录 一 概要 基于单片机温湿度PM2.5报警设置系统概要主要涵盖了系统的整体设计思路、硬件组成、软件实现以及报警功能等关键方…

HCIP综合实验1

一.实验拓扑 二.实验要求 1、R5为ISP&#xff0c;只能进行IP地址配置&#xff0c;其所有地址均配为公有Ip地址; 2、 R1和R5间使用PPP的PAP认证&#xff0c;R5为主认证方&#xff1b; R2与R5之间使用PPP的CHAP认证&#xff0c;R5为主认证方; R3与R5之间使用HDLC封装; 3、R2、R3…

高密度集成可伸缩无机薄膜晶体管

引言 与有机晶体管相比&#xff0c;无机半导体晶体管具有优越的性能和可靠性。然而&#xff0c;由于其脆性&#xff0c;不利于制造可拉伸的电子产品。由于这一缺点&#xff0c;它们大多被放置在不可拉伸的部件上&#xff0c;以避免机械应变&#xff0c;加重连接这些刚性部件的…