·

Spark and Big Data Application

发布时间:2024-07-23 10:57:13阅读量:390
学术文章
·
介绍文
转载请注明来源

Spark and Big Data Application

Digital data is growing fast in volume courtesy of the innovations that are taking place in the digital world. The data quantities are huge and are from different sources, including smartphones and social networks, which the relational databases and the analytical techniques are incapable of storing or processing. These huge amounts of data get the name Big Data from this characteristic. Traditional machine learning cannot handle the volume and velocity of the data and it therefore requires the right system and programming model to complete the analysis. Big data comes in different forms, and their distribution and analysis are possible through data frameworks like Hadoop and Spark. Large data analytics can also benefit from the use of different types of Artificial Intelligence (AI) like Machine Learning for results. The use of big data tools and AI techniques has changed the phase of data analysis. This paper will examine a big data computing model based on Spark and the application of big data. Additionally, it will highlight big data platforms.

Big Data Platforms

Several platforms form big data, including batch processing, real-time processing and interactive analytics. In batch processing, there are extensive computations that also require time for data processing. The most common form of batch processing is Apache Hadoop. The platform is scalable, cost-effective, flexible and tolerates faults involved in the processing of big data. There are also other Hadoop Platforms like the Hadoop Distributed File System (HDFS), Yet Another Resource Negotiator (YARN) and MapReduce (Rahmani et al., 2021). These platforms can also operate across the big data value chain, which works from aggregating, storing, processing and managing big data. When it comes to stream processing platforms, Apache Spark and Storm are the most popular platforms. Such applications used in stream processing play a big part in weather and transport systems. It is possible to access datasets and use them for various operations remotely. The users can also access the system directly and interact with the data. An example of an interactive analytic platform is the Apache Drill.

Apache Spark Big Data Computing

It is a text-based framework whose origin traces back to the University of Berkeley back in 2009 and has since become a powerful tool capable of macro processing, unlike other tools. The core is the basis of the features that are found in Spark. Spark has three application programming interfaces (APIs), namely Java, Python and Scala, and its core is an API location that acts as a definition for the RDD dataset found in its original programming. Apache Spark supports the building and running of fast and high-level sophisticated applications on Hadoop. Spark engine is responsible for supporting tools that can support SQL, for instance, queries, stream data applications, and manage complex analytics like machine learning and graph algorithms. There are several big data ecosystems based on the Spark platform, including Spark runtime, spark SQL, graph X, Spark streaming and even the MLlib-Algorithm library (Chen, 2021).

The analysis of spark runtime requires the determination of functionality, for instance, the scheduling of tasks and the management of memory. In instances that involve data transmission inside Spark using RDD structure, the idea is for the determination of the core logic data information of Spark. The first will involve the division of data into many subsets and transmitting the subsets to nodes in the cluster to help in processing. The second steps lead to result protection after the calculation. The protection it is possible to have the same file content as the results of the calculation and it also becomes easier to back up the file content. The final step happens only in the case where calculation errors occur at any of the subsets, which will involve the resorting of the subset to do away with the errors.

Shark is where Spark SQL is developed. Shark employs the concepts from HQL parsing, logical execution plan translation, and execution plan optimization in hive to achieve hive compatibility. One could argue that the spark job merely replaces the physical execution plan. Additionally, it is dependent on the past Metastore and Hive series. When it comes to historical compatibility, Spark SQL solely uses the HQL parser, history meta store, and Hive server. Stated differently, spark SQL takes over since HQL is parsed into an abstract syntax tree (AST). The catalyst is in charge of creating and optimizing the execution plan. It is considerably easier to build in Scala thanks to its functional language capabilities, such as pattern matching.

Among the many benefits of Spark, SQL integration is the most noteworthy. Spark SPL seamlessly combines spark programs and SQL queries with spark SQL. Users may query structured data in the spark program using conventional data frame APIs or SQL. It is compatible with R, python scala and Java by employing knowledgeable API to work with the data. This benefit can increase the effectiveness of our research. Running SQL or hiveql queries on pre-existing warehouses with Hive integration is the second benefit. Unified data access offers a third benefit since it connects to any data source in an identical manner. Access to a range of data sources such as hive Avro parquet ORC JSON and JDBC can be facilitated by using data frames and SQL even adding data from different sources is possible. The final advantage comes in the form of unified data access, which allows for standard data connections via JDBC or ODBC connections. Server mode supports industry-standard JDBC and ODBC connections for business intelligence products.

When the calculating algorithm is analyzed using the Mllib algorithm library, it is shown to have a high level of complexity. Iterative computations need a lot of CPU power because all calculations must be stored on disk in order to wait for task processing to begin.

A portion of the work can be completed entirely in memory using the Spark platform. Because the iterative calculation is becoming more efficient, the corresponding iteration portion of the calculation task is moved straight to memory. It can also achieve disk and network operation under specific conditions. In conclusion, Spark offers even more impressive benefits for iterative computing, which can be used as a platform for distributed machine learning. Using a collaborative filtering algorithm is meant to figure out the concept before sharing the idea with the users.

There are steps necessary in the algorithm application, and the description is as follows. System filtering comes first. After choosing the people who share similar interests, the objects are chosen and categorized based on their preferences, creating a new set or sequence. Users in this process can be thought of as neighbors, and it is best to arrange and make use of related users in order to ascertain the best execution strategy. Secondly, cooperative filtering. The integration of user preferences is the essential component of the final suggestion, which is completed by means of the phases of user preferences gathering, similarity analysis, and recommendation based on the computation results. To accurately gather data behavior, the first consideration is choosing a user system and then organizing it based on user behavior and subsequently processing the data before issuing recommendations on what users may like provided their preferences.

The spark system is responsible for expanding spark data and also divides Spark streaming data depending on the time mode and eventually forms RDD. When processing streaming data that has small intervals, there will likely be a processing delay which leads to the time processing system. The advantage of using spark streaming includes its response to a fault, which, as it is, has a strong fault tolerance that makes it capable of handling errors and recovering data. Another benefit of using it is its ability to connect to relevant spark modules. One of its added advantages, apart from completing the flow of data, is the ability to deal with complexity.

Another key component of Spark is Graph X, whose foundation is on the large -scale graph computing with the ability to compute large graph data processing in Spark with the aid of Graph X properties. In the event of analysis of graph x, it is verifiable that it could be a source of rich data operators like the core and optimization operators. Graph X also has the capability of meeting graph operations of several distributed clusters and has enough API interfaces. According to Chen (2021), Graph X plays a critical role in Spark as it improves its data absorption and scale.

Applications of Big Data

Medical Application

Healthcare facilities and medical organizations use big data to anticipate potential health risks and prevent them. The healthcare system acquires large benefits from large-scale data analysis. For example, analytics has made it possible to diagnose diseases in their early stages, like cases of breast cancer. It has also supported the processing of medical images and medical signals for providing high-quality diagnosis and monitoring of patient symptoms. Additionally, tracking chronic diseases such as diabetes has been easier. Other uses are preventing the incidence of contagious diseases, education through social networks, genetic data analysis and personalized medicine. Data types include biomedical data, web data, data from different electronic health records (EHRs), hospital information systems (HISs), and omics data, which includes genomics, transcriptomics, proteomics, and pharmacogenomics. Rich data, such as demographics, test results, diagnoses, and personal information, are included in the EHR and HIS data (Ismail Ebada et al., 2022). Consequently, the significance of health data analysis has been taken into consideration, which has led to knowledge for scholars to control, manage and process big data courtesy of tools like Apache Spark.

Apache Spark's memory calculations enable rapid processing of large healthcare datasets, both unstructured and structured. It can analyze data faster than typical map reduction approaches, approximately 100 times faster. Spark's lambda architecture enables both real-time and batch processing. The tool analyzes streaming healthcare data and includes a library for handling Twitter data streams. Several messages relate to health, and communication is possible through the use of social networks. Twitter, for instance, provides real-time public health statistics that are less expensive to get since real-time disease surveillance entirely depends on mining the information from Twitter API methods (Nair et al., 2018). Spark's machine learning package, MLlib, can help create decision trees for illness prediction. Spark analyzes large amounts of data as it streams. It outperforms Hadoop in streaming due to its memory-based processing capabilities. Data and intermediate outcomes are kept in memory, which helps avoid delays that come in the input-output stage of data that has to move back and forth on the hard disk.

Spark uses resilient distributed datasets (RDDs), which are immutable object collections stored in memory and partitioned among nodes for parallel calculations. Spark is a batch-processing engine that can handle streaming data from several sources, including Twitter. The Spark engine handles small batches of the incoming data streams after division. Discretized streams (DStreams) are a series of RDDs that symbolize a continuous data stream in Spark streaming. The Spark engine computes the fundamental RDD transformations that result from operations on DStreams. Popular learning methods that aid in solving machine learning problems include collaborative filtering, clustering, and classification found in MLlib.

Natural Disaster Prediction and Management

The use of big data is a crucial part of responding to natural disasters. Some of the applications of the tools are in preventing disasters, providing relief in times of disaster and managing disasters in times of emergency. In the era of the Internet of Things (IoT), the complexity of natural disaster management has become simpler. AI Spark Big Model has enabled a new way of thinking, technical system and innovative ability to deal with complex disaster information, resulting in an effective way of dealing with comprehensive disaster prevention and reduction. Big data has strategic significance and, together with its core values, is mainly composed of three concepts. The first is on the level of strategic thinking which defines the point when the disaster problem is beyond the ability of contemporary scientific understanding. Such problems are high-frequency and high-intensity disasters and go beyond the ones that the world defines as basic disasters.

Additionally, comprehensive disaster reduction goes beyond the national strategy. The level of innovation and development of information science and technology in big data refers to the second concept that defines a situation where disaster big data results in challenges in traditional information science and technology. At this point, there is an acceleration in the formation of technology systems and new theories, which have complex information serving as the core. The third concept defines the level of social innovation and development. Here, disaster big data is the core capability of ensuring that the world has Internet and is smartly secure and can reduce disaster using these smart strategies.

The ability to efficiently monitor disaster information, standardize data, predict abnormalities and provide early warnings of signs of disaster, the existing forms assessing disaster risks and all kinds of disasters form disaster cloud computing (Chen, 2021). Since it deals with early warning, analyzes how to manage emergencies, provides emergency information services and analyses the calculation of other models, it is capable of building a large distributed ring in the way of shared infrastructure. It serves the major purpose of providing safe and faster disaster cloud data processing and predicts and issues an early warning. It also provides a disaster risk assessment model, emergency management service model and other related services.

Face Recognition

Face recognition is another application of big data analytics tools like Apache Spark. Businesses are able to use big data analysis systems to detect faces, recognize the faces, recognize gender, age, facial expressions and gazes of people watching the machines advertising and other digital signs that originate from face recognition technology. The information that big data processing technology collects becomes an instrumental part of the analysis of customer preferences. In the end, it helps provide customer-friendly services according to their gender, age and emotional state. It generally facilitates the connection between people and things, hence capable of making recommendations on the advertisements and messages that certain individuals prefer. It is also useful in the non-manual and automatic uses aimed at recognition of face, eyes and any other information that could help in predicting subsequent operations.

Conclusion

The paper addresses various issues, starting with the definition of big data and analyzing the types of platforms that form big data. It also summarizes the applications in Apache Spark by defining Spark SQL, graph x, Mllib algorithm library and Spark streaming and how they operate. It, however, still needs to be determined which platforms are stable, and it is therefore advisable to use the tools that have undergone tests and proven their scalability, flexibility and stability. Additionally, the essay introduces the reader to the areas where big data play crucial roles. The areas include its use in the medical field in the prediction, analysis, prevention and management of various forms of diseases. Other areas include its use in the management and prediction of natural disasters and its use in face recognition, among other areas. Technology is still evolving, and big data modelling could still use some improvements like having an enhanced real-time data processing which other tools like Apache Spark have already made possible courtesy of Twitter APIs. Most industries are also moving toward using big data processing systems that are accurate and precise to enhance their economic and social benefits. Batch processing, which is the most common big data processing, needs to meet the frequency of data processing requirements that these industries need.

References

  1. Chen, S. (2021). Research on Big Data Computing Model based on Spark and Big Data Applications. Journal of Physics: Conference Series, 2082(1), 012017. https://doi.org/10.1088/1742-6596/2082/1/012017
  2. Ismail Ebada, A., Elhenawy, I., Jeong, C.-W., Nam, Y., Elbakry, H., & Abdelrazek, S. (2022). Applying Apache Spark on Streaming Big Data for Health Status Prediction. Computers, Materials & Continua, 70(2), 3511–3527. https://doi.org/10.32604/cmc.2022.019458
  3. Nair, L. R., Shetty, S. D., & Shetty, S. D. (2018). Applying Spark based machine learning model on streaming big data for health status prediction. Computers & Electrical Engineering, 65, 393–399. https://doi.org/10.1016/j.compeleceng.2017.03.009
  4. Rahmani, A. M., Azhir, E., Ali, S., Mohammadi, M., Ahmed, O. H., Yassin Ghafour, M., Hasan Ahmed, S., & Hosseinzadeh, M. (2021). Artificial intelligence approaches and mechanisms for big data analytics: a systematic study. PeerJ Computer Science, 7, e488. https://doi.org/10.7717/peerj-cs.488
评论区

暂无评论,来发布第一条评论吧!

弦圈热门内容

10月底至11月初,弦圈功能更新:上传附件

最近我们将给“写帖子”添加一个新的功能:上传附件。也就是说,以后在弦圈用户可以上传自己写的pdf版或者doc版的文章,或者是可以分享自己收藏的电子书。至此,弦圈将同时兼顾为一个资源型网站,大家可以在弦圈的帖子中寻找相关的资源(如电子书、notes等)进行下载。附件的下载可以设置条件,如回复后才可下载,支付金币可下载等。之后为了更好的体验,我们计划引入智力(经验)和货币系统。货币暂分为金币和弦币,其中金币为免费的,所谓“书中自有黄金屋”,签到可获得智力值,而智力值又能产生金币😇。最后,如果上传的附件存在侵权等问题,请联系删除。注意,最新帖子泛函分析教材Functional Analysis Notes(2011)为测试贴(下图为测试画面),新功能将马上上线😃。11月2日:新功能已于昨日更新完成,欢迎尝试使用!😃

如果我看数学看得很慢,这没问题吗?

我在一所知名的数学学院读数学本科,今年是最后一年。然而我发现一件事情,那就是我好像看数学的速度要比班上其他同学慢。比如,无论我尝试多少遍,我似乎都是班上最后做完作业的人,并且我很少有空余时间进行课外阅读。你觉得有哪些建议或者技巧是我可以尝试的?或者说为了节省时间,我是不是应该跳过细节?回答1:提问和给出的信息有些模糊,可能无法给出令人满意和有意义的答案。但我仍然会尝试给出一个答案:我想我们每个人都知道数学中的这些短语,如“easy to see”或者类似的词组,他们能占据一个人数小时注意力,并且显然会导致读完一篇文章所需要花费的时间,比理所当然的要更长。因此如果你为此付出了更多努力,而你的同学们却没有,那读得慢确实没什么问题。还有当你第一次阅读文本的时候,你是否会尝试理解每一个证明中的每一处细节?我非常肯定这不是你每个同龄人都能做到的。并且我发现有几种不同的“类型”。比如说我在第一次看时,往往需要先有个大概的了解,然后再深入理解更为复杂的证明和细节。我同时喜欢多次反复阅读一个文本,因为我记性不好,这或许会让我重复一些东西,但当然也意味着我第一次会看得很快,但也很肤浅(所以我先从鸟的视角 ...

为什么无限求和需要被有意义的?

我的提问:例如单位分解(partition of unity)中的求和以及抽象代数中的多项式表达式。回答:拥有无限多项的求和(或者说更加正式的“级数”)需要一些额外的条件来保证他们“表现良好”("well behaved")。否则你可能得到像以下这样的悖论:$$\begin{align} &S = 1 + 1 + 1 + \dots \\ &\Rightarrow 2S = 2 + 2 + 2 + \dots \\ &\Rightarrow 2S = (1+1) + (1+1) + (1+1) + \dots \\ &\Rightarrow 2S = 1 + 1 + 1 + \dots \\ &\Rightarrow 2S=S \\ &\Rightarrow S = 0 \end{align}$$一般地,额外的条件包含,要求除了有限数量的项都为$0$(数学简称中的“几乎所有”)或者收敛条件来确保求和有一个极限值。本问题问于2020年1月22号,当时我在读高三,提问的水平非常差😅,跟Peter Scholze这种高中就懂谱序列的没得比🙃。

范畴中的态射一定得保持结构吗?我在教材中找到了一些不一样的

我的提问:众所周知,范畴中对象之间的态射都是保持结构的。但是在一本教材中,我发现它说态射一般是保持结构的。这是否意味着存在不保持结构的态射?回答1:一个范畴不需要非得由带有某些额外结构的集合与保持这个结构的映射构成。不是这种类型的范畴的例子有:给定任意一个群$G$,我们可以构造一个范畴,它由一个对象$*$和每个$g\in G$的一个态射$\varphi_g\colon *\to *$组成。这里,态射的复合通过群运算来定义,并且$\operatorname{id}_* = \varphi_{e}$对于单位元$e\in G$。给定一个偏序集$(P,\le)$,我们可以构造一个范畴,它由对象集$P$和每个满足$x\le y$的$x,y\in P$有且仅有一个的态射$x\to y$组成。拓扑空间的同伦范畴,它的对象都是拓扑空间,每个态射$X\to Y$是一个连续映射$f\colon X\to Y$的同伦群$[f]$。回答2:我认为问题出在这里众所周知,范畴中对象之间的态射都是保持结构的。事实并非如此。范畴这个概念推广了“带有结构的集合和保持结构的函数”,例如群和同态,或者拓扑空间和连续映射。但 ...

我翻译并整理了一些MathStackExchange的问题和回答

对于数学老手而言,阅读全英文数学甚至是全法语数学,都是可以做到的。但是对于数学萌新而言,阅读全英文的数学内容,可能会比较吃力,也需要花费更多的时间来进行阅读和理解。然而对于做数学的人而言,不懂英文就意味着会有大量优质的英文数学资源无法享用。国外比较有名的数学论坛包括MathStackExchange 与 MathOverflow,都拥有大量优秀的问题以及十分优质的回答,这往往能帮助你解决学习过程中遇到的难题。所以,我觉得可以翻译一些MathStackexchange与MathOverflow的优质内容,让更多的国内的数学爱好者能够接触到优秀的英文数学资源。目前我已翻译,并重新整理以下内容,中英对照(切换语言可见):如何构建一个比复数域ℂ还要大的域?ℝ的有限域扩张是ℝ或者同构于ℂ幂零理想层的局部截面是什么样的?我在哪可以找到一个数学笔友?范畴中的态射一定得保持结构吗?我在教材中找到了一些不一样的阿基米德性质的乘法形式如果我看数学看得很慢,这没问题吗?仿射概形上的概形什么时候仿射?如果两个对象的余极限同构,那么这两个对象同构?正弦函数的幂级数展开是否是柯西序列?任意一个范畴之间的本质满射都 ...

关于平方可积空间的一些疑问

勒让德多项式是平方可积空间的完备正交基,也是该空间的绍德尔基,即该空间的任意一个元素,可以由其唯一的表示。而勒让德多项式是由x,x^2,x^3....通过施密特正交化原理得到的,两者张成的空间相等都在平方可积空间中稠密,那么我想问的是x,x^2,x^3....是该空间的绍德尔基,即平方可积空间的任意一个元素可由x,x^2,x^3....唯一表示么

抽象代数中如何执行归纳法?

我的提问:我无法理解在这个证明中,归纳法这个步骤是如何进行的。有人能帮帮我吗?感谢!回答:令$n = deg B$。他们通过对$m = deg A$做归纳法来证明那个陈述。基本情况是$m < n$。如果$m \geq n$,然后他们找到另一个多项式$A'$,在这种情况下,$A' = A - B a_m X^{m - n}$,并且它有比$m$更小的阶数。所以我们可以通过归纳假设来处理它。$A′$的商和余数表达式是用于找到$A$的。我想有两件事你可能会觉得困扰,以及为什么你没有认出归纳法。首先,基本情况不仅仅是一种情况,而是一堆情况。这里请注意,这是基本的:证明中的归纳步骤仅适用于$m\geq n$。同时注意,在这种情况下,证明$m=1$的工作量并不比证明$m<n$小:对于所有这些情况,这都是一行证明。你可能会觉得困扰的第二件事是,我们不仅对$m-1$使用归纳假设,对任何阶数严格小于$m$的多项式也使用归纳假设。这被称为完全归纳法或强归纳法:在归纳步骤中,你假设的是,命题不多于$m-1$时都是真的,而不仅仅是$m-1$。这在维基百科的“归纳法”页面上得到了很好的解释。

如何理解$\mathbb{Q}_{p}(p^{1/p^{\infty}})$?

我的提问:众所周知$\mathbb{Q}_{p}(p^{1/p^{\infty}})$被定义为$\bigcup_{n>0} \mathbb{Q}_{p}(p^{1/p^{n}})$,意思是邻接所有$p$的$p$幂根($p$-power roots of $p$)到混合特征域$\mathbb{Q}_{p}$。然而,我不太懂这个符号的意思$\mathbb{Q}_{p}(p^{1/p^{n}})$。这是如何联系到$p$的$p$幂根的?为何在这个记号中,$p$的幂是$1/p^{n}$?我认为$\mathbb{Q}_{p}(p^{1/p^{n}})$是$\mathbb Q_p$的一个割圆扩张,其中$p^{1/p^{n}}$是$n$次单位本原根(primitive $n$th root of unity)。但是似乎这说不通。并且我在另一个回答中看到$\mathbb{Q}_{p}(p^{1/p^{n}})$是一个分歧扩张(ramified extension)。谁能告诉我在哪里可以了解$\mathbb{Q}_{p}(p^{1/p^{n}})$?回答1:根据定义,$\Bbb Q_p(p^{1/p ...

可代表层的满射性

我的提问:令$S$为一个基概形,并令$(Sch/S)_{fppf}$为一个大fppf景。令$U$为一个$S$上的概形。假设存在一个满射态射$\Phi_{U}:U\rightarrow U$。那么我们能证明导出的层态射$h_{U}\rightarrow h_{U}$局部满射的?这看起来是错误的。注意到$h_{U}={\rm{Hom}}(-,U)$是一个可代表层。一个$(Sch/S)_{fppf}$上的层映射$F\rightarrow G$是局部满射的,如果对每个概形$U\in{\rm{Ob}}((Sch/S)_{fppf})$和每个$s\in G(U)$,都存在一个覆盖$\{U_{i}\rightarrow U\}_{i\in I}$,使得对所有$i$,$s|_{U_{i}}$在$F(U_{i})\rightarrow G(U_{i})$的像中。回答:令$S:={\rm Spec}(k)$为一个域,并且令$U={\rm Spec}(k[t]/t^2)$。环$k[t]/t^2$是一个$k$-代数,并且存在一个$k$-代数映射$k[t]/t^2\to t$,其将$t$打到$0$,所以我们得到 ...

任意一个范畴之间的本质满射都是一个满态射吗?

我的提问:令$\cal{C},\cal{D}$为范畴(或者栈)。令$F:\cal{C}\rightarrow\cal{D}$是一个本质满射的函子,即在对象同构类上满射。然后$F$是小范畴(或者栈)范畴中的一个满态射吗?回答:不是。例如,任何一个对象的范畴之间的函子是本质满射的,但是如果$M_1, M_2$是两个非零幺半群,那么一个直和项的包含映射$M_1 \to M_1 \oplus M_2$,看成是两个单对象范畴间的一个函子,不是一个范畴的满态射。不过记住,“小范畴范畴中的满态射”由于多种原因,在任何特定应用中,都显然不是“正确”的概念。它抛弃了自然变换,所以你忽略了这样一个事实,即你其中在2-范畴里操作;并且在任何特定情况下,你可能需要各种“满态射”的概念。

10.27 弦圈问题分析以及改进计划

最近有不少对弦圈感兴趣的爱好者,在弦圈注册了账号,也有人参与了互动。对此,我在这感谢各位的支持和认可!😃不过经过这段时间,用户注册后的表现,也透露出目前弦圈存在的很多问题。首当其冲的就是首页,默认显示最新内容,用时间顺序排序,意味着大家在首页往往无法看到有趣的内容,也可能找不到他想看的内容。这也导致弦圈中优秀的内容被埋没。因此,针对这个问题,我自己设计了一个简单的热度算法来计算“热度”,然后通过“热度”来排序首页的热门内容。旧的热门内容就是单纯的通过阅读量排序,没有热度随着时间衰减的现象,这也意味着新内容往往容易被旧内容排挤掉。有了更好的热度算法,我就可以将打开首页默认显示最新内容,改为默认显示热门内容了😇。接着就是中英文混合的问题,这个首页已经解决了,首页看到的内容都会把其他语言的给过滤掉。但是圈子内的话,我没有强行设置只有一种语言,因为不太想一些优秀的英文内容被埋没。我有点想参考推特的做法:热门内容推荐的大多数都是一种语言(如中文),只有一两个是其他语言(如英文)。或者说还有一种方案:热门内容全是同一种语言,再增加一个选项”全部“,即查看圈子全部内容。至于数学圈首页,那些数学分支的 ...