Warning: file_get_contents(/data/phpspider/zhask/data//catemap/9/loops/2.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
R:循环和可视化;“运行时间”;在R_R_Loops_Runtime_Data Visualization_Data Manipulation - Fatal编程技术网

R:循环和可视化;“运行时间”;在R

R:循环和可视化;“运行时间”;在R,r,loops,runtime,data-visualization,data-manipulation,R,Loops,Runtime,Data Visualization,Data Manipulation,我正在使用R编程语言。我想学习如何随着数据大小的增加测量和绘制差异过程的运行时间 我发现以前的stackoverflow帖子回答了一个类似的问题: 看起来R中的“微基准”库应该能够完成这项任务 假设我模拟以下数据: #load libraries library(microbenchmark) library(dplyr) library(ggplot2) library(Rtsne) library(cluster) library(dbscan) library(plotly) #sim

我正在使用R编程语言。我想学习如何随着数据大小的增加测量和绘制差异过程的运行时间

我发现以前的stackoverflow帖子回答了一个类似的问题:

看起来R中的“微基准”库应该能够完成这项任务

假设我模拟以下数据:

#load libraries

library(microbenchmark)
library(dplyr)
library(ggplot2)
library(Rtsne)
library(cluster)
library(dbscan)
library(plotly)

#simulate data

var_1 <- rnorm(1000,1,4)
var_2<-rnorm(1000,10,5)
var_3 <- sample( LETTERS[1:4], 1000, replace=TRUE, prob=c(0.1, 0.2, 0.65, 0.05) )
var_4 <- sample( LETTERS[1:2], 1000, replace=TRUE, prob=c(0.4, 0.6) )


#put them into a data frame called "f"
f <- data.frame(var_1, var_2, var_3, var_4)

#declare var_3 and response_variable as factors
f$var_3 = as.factor(f$var_3)
f$var_4 = as.factor(f$var_4)

#add id
f$ID <- seq_along(f[,1])
#加载库
图书馆(微基准)
图书馆(dplyr)
图书馆(GG2)
图书馆(Rtsne)
图书馆(群集)
图书馆(dbscan)
图书馆(绘本)
#模拟数据

var_1创建一个函数,该函数执行分析的所有步骤,并将其传递到
microbenchmark
。在伪代码中,与

runAnalysis <- function(x, size) {
  x <- x[1:size, ]
  # forther steps of the analysis
}

xy <- microbenchmark(
  subset_5 = runAnalysis(x = f, size = 5),
  subset_50 = runAnalysis(x = f, size = 50),
  times = 1
)
runAnalysis
procedure_1_part_1 <- microbenchmark(daisy(f[,-5],
                    metric = "gower"))

procedure_1_part_2 <-  microbenchmark(as.matrix(gower_dist))
runAnalysis <- function(x, size) {
  x <- x[1:size, ]
  # forther steps of the analysis
}

xy <- microbenchmark(
  subset_5 = runAnalysis(x = f, size = 5),
  subset_50 = runAnalysis(x = f, size = 50),
  times = 1
)