Software Component Block Diagram
RPM(Resource Power Manager)是高通MSM平台另外加的一块芯片,虽然与AP芯片打包在一起,但其是一个独立的ARM Core。之所以加这个东西,就是要控制整个电源相关的shared resources,比如ldo,clock。负责与SMP,MPM交互进入睡眠或者唤醒整个系统。
以下是高通文档中对各个功能模块的说明。
- Kernel – DALSys-based lightweight kernel
- RPM handler – RPM handler abstracts the RPM message protocol away
from other software - Drivers – Drivers for each of the resources supported by the RPM will
register with RPM handler to request notification when requests are
received for the resource which the driver controls - NPA – A driver may use the Node Power Architecture (NPA) to represent
resources controlled by the driver - Clock driver – RPM clock driver directly handles aggregating requests
from each of the masters for any of the systemwide clock resources
controlled by the RPM - Bus arbitration driver – RPM bus arbiter driver takes bus arbiter settings
as requests from the different masters in the system and aggregates them
to represent the frequency-independent system settings - PMIC – RPM PMIC driver directly aggregates requests from each of the
masters for any of the systemwide PMIC resources controlled by the RPM - Watch Dog driver – Watch Dog driver is a fail-safe for incorrect or stuck
code - MPM driver – Used to program the MPM hardware block during
systemwide sleep - RPM message driver – RPM message protocol driver abstracts the RPM
message protocol away from other subsystem software
Messaging Masters
与RPM通过shared memory region交互进行dynamic and static resource/power management的可以有很多种。
这个可以查看smd_type.h中的smd_channel_type。但目前看只有AP,MODEM,RIVA,TZ与RPM有交互,这个可以看message_ram_malloc()函数中的设置。
其实也可以间接从rpm_config.c文件中的SystempData temp_config_data这个变量中看出来到底有几个部分是与RPM进行交互的。
RPM Initialization
在main.c文件中会逐个调用init_fcns[]变量中的函数进行初始化。当然也包括上面的资源的初始化。
const init_fcn init_fcns[] ={ populate_image_header, npa_init,#if (!defined(MSM8909_STUBS) ) railway_init_v2,#endif PlatformInfo_Init, pm_init, acc_init,#if (!defined(MSM8909_STUBS) ) railway_init_early_proxy_votes, #endif (init_fcn)Clock_Init, __init_stack_chk_guard, ddr_init, smem_init, init_smdl, version_init, rpmserver_init, rpm_server_init_done, railway_init_proxies_and_pins,#if (!defined(MSM8909_STUBS) ) rbcpr_init,#endif svs_init, vmpm_init, sleep_init,#if (!defined(MSM8909_STUBS) ) QDSSInit, #endif exceptions_enable, swevent_qdss_init, icb_init,#if (!defined(MSM8909_STUBS) ) debug_init, system_db_init, zqcal_task_init,#endif rpm_settling_timer_init, gpio_toggle_init, rpm_set_changer_common_init,}
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
- 29
- 30
- 31
- 32
- 33
- 34
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- 44
- 45
- 46
- 47
- 48
- 49
- pm_init : LDO等资源的注册,然后接收rpm_message。接收rpm_message的部分高通代码没有给,所以看不到,但从rpm log来看,是有接收处理并反馈的过程的。
107.764902: rpm_message_received (master: "APSS") (message id: 723)107.764908: rpm_svs (mode: RPM_SVS_FAST) (reason: imminent processing)107.764924: rpm_svs (mode: RPM_SVS_FAST) (reason: imminent processing)107.764939: rpm_process_request (master: "APSS") (resource type: ldoa) (id: 14)107.764942: rpm_xlate_request (resource type: ldoa) (resource id: 14)107.764946: rpm_apply_request (resource type: ldoa) (resource id: 14)107.765033: rpm_send_message_response (master: "APSS")
ldoa对应的resource type为RPM_LDO_A_REQ。这个在pm_rom_device_init()里的pm_rpm_ldo_register_resources(RPM_LDO_A_REQ, num_of_ldoa); 这里被注册,所以看进去最后xlate和apply最终都会被pm_rpm_ldo_translation()和pm_rpm_ldo_apply进行处理。pm_rpm_ldo_tranlation()读取kvp内容,pm_rpm_ldo_apply最终把request的内容设置上去。kernel端在msm-pm8916-rpm-regulator.dtsi文件中定如下ldo内容
2. (init_fcn)Clock_Init:Clock资源的注册,这个过程和上面的也差不多
3. rpmserver_init:启动接收message的进程
比如要设置LDO3的电压和电流,kvp的内容如下:
{ “req\0” : { {“rsrc” : “ldo\0”} {“id” : 3} {“set” : 0} {“data” : { {“uv\0\0” : 1100000} {“mA\0\0” : 130} } } }}
RPM log里边显示是APSS,这个在kernel里边也是有设置的,在msm-pm8916-rpm-regulator.dtsi里的
rpm-regulator-ldoa14 { compatible = "qcom,rpm-smd-regulator-resource"; qcom,resource-name = "ldoa"; qcom,resource-id = <14>; qcom,regulator-type = <0>; qcom,hpm-min-load = <5000>; status = "disabled"; regulator-l14 { compatible = "qcom,rpm-smd-regulator"; regulator-name = "8916_l14"; qcom,set = <3>; status = "disabled"; }; };
MSM的retulator相关的驱动会读取这个,如果有需要发请求的话,会根据这个dtsi的内容,组织kvp,然后通过rpm_request通道发过去。
rpm_request通道在rpm里边打开的地方是在rpm_handler.cpp里边的Hander::init()函数
void Handler::init(){ smdlPort_ = smdl_open("rpm_requests", rpm->ees[ee_].edge, SMDL_OPEN_FLAGS_MODE_PACKET, rpm->ees[ee_].smd_fifo_sz, (smdl_callback_t)rpm_smd_handler, this);}
在对应的kernel端也有打开,可以参考msm8916.dtsi文件里边的设置。
rpm_bus: qcom,rpm-smd { compatible = "qcom,rpm-smd"; rpm-channel-name = "rpm_requests"; rpm-channel-type = <15>; };
以下是注册资源的函数和资源的类型:
所有的资源都通过rpm_register_resource()函数注册。这些资源包括clock, ldo等。具体可以看下面的类型定义。
typedef enum{ RPM_TEST_REQ = 0x74736574, RPM_CLOCK_0_REQ = 0x306b6c63, RPM_CLOCK_1_REQ = 0x316b6c63, RPM_CLOCK_2_REQ = 0x326b6c63, RPM_CLOCK_QPIC_REQ = 0x63697071, RPM_BUS_SLAVE_REQ = 0x766c7362, RPM_BUS_MASTER_REQ = 0x73616d62, RPM_BUS_SPDM_CLK_REQ = 0x63707362, RPM_BUS_MASTER_LATENCY_REQ = 0x74616c62, RPM_SMPS_A_REQ = 0x61706D73, RPM_LDO_A_REQ = 0x616F646C, RPM_NCP_A_REQ = 0x6170636E, RPM_VS_A_REQ = 0x617376, RPM_CLK_BUFFER_A_REQ = 0x616B6C63, RPM_BOOST_A_REQ = 0x61747362, RPM_SMPS_B_REQ = 0x62706D73, RPM_LDO_B_REQ = 0x626F646C, RPM_NCP_B_REQ = 0x6270636E, RPM_VS_B_REQ = 0x627376, RPM_CLK_BUFFER_B_REQ = 0x626B6C63, RPM_BOOST_B_REQ = 0x62747362, RPM_SWEVENT_REQ = 0x76657773, RPM_OCMEM_POWER_REQ = 0x706d636f, RPM_RBCPR_REQ = 0x727063, RPM_GPIO_TOGGLE_REQ = 0x6F697067, } rpm_resource_type;
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
RPM Resouce Registration
- rpm_register_resource():
Gpio_toggle.c (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\common): rpm_register_resource(RPM_GPIO_TOGGLE_REQ, 3, sizeof(gpio_toggle_inrep), gpio_toggle_xlate, gpio_toggle_apply, 0);Icb_rpm_spdm_req.c (x:\j3_ctc\nhlos\rpm_proc\core\buses\icb\src\common): rpm_register_resource( RPM_BUS_SPDM_CLK_REQ,Ocmem_resource.c (x:\j3_ctc\nhlos\rpm_proc\core\power\ocmem\src): rpm_register_resource(RPM_OCMEM_POWER_REQ, 3, sizeof(rpm_ocmem_vote_int_rep), rpm_ocmem_xlate, rpm_ocmem_apply, 0);Pm_rpm_boost_trans_apply.c (x:\j3_ctc\nhlos\rpm_proc\core\systemdrivers\pmic\npa\src\rpm): rpm_register_resource(resourceType, num_npa_resources + 1 , sizeof(pm_npa_boost_int_rep), pm_rpm_boost_translation, pm_rpm_boost_apply, (void *)boost_data);Pm_rpm_clk_buffer_trans_apply.c (x:\j3_ctc\nhlos\rpm_proc\core\systemdrivers\pmic\npa\src\rpm): rpm_register_resource(resourceType, num_npa_resources + 1, Pm_rpm_ldo_trans_apply.c (x:\j3_ctc\nhlos\rpm_proc\core\systemdrivers\pmic\npa\src\rpm): rpm_register_resource(resourceType, num_npa_resources + 1, sizeof(pm_npa_ldo_int_rep), pm_rpm_ldo_translation, pm_rpm_ldo_apply, (void *)ldo_data);Pm_rpm_smps_trans_apply.c (x:\j3_ctc\nhlos\rpm_proc\core\systemdrivers\pmic\npa\src\rpm): rpm_register_resource(resourceType, num_npa_resources + 1, sizeof(pm_npa_smps_int_rep), pm_rpm_smps_translation, pm_rpm_smps_apply, (void *)smps_data);Pm_rpm_vs_trans_apply.c (x:\j3_ctc\nhlos\rpm_proc\core\systemdrivers\pmic\npa\src\rpm): rpm_register_resource(resourceType, num_npa_resources + 1, sizeof(pm_npa_vs_int_rep), pm_rpm_vs_translation, pm_rpm_vs_apply, (void *)vs_data); Rpmserver.cpp (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\server):void rpm_register_resourceRpmserver.h (x:\j3_ctc\nhlos\rpm_proc\core\api\power):void rpm_register_resourceRpm_npa.cpp (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\server): rpm_register_resource(resource, num_npa_resources, sizeof(npa_request_data_t), rpm_npa_xlate, rpm_npa_apply, adapter);Rpm_npa.cpp (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\server): rpm_register_resource(resource, num_npa_resources, sizeof(npa_request_data_t), rpm_npa_xlate, rpm_npa_apply, adapter);Rpm_test_resource.cpp (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\server): rpm_register_resource(RPM_TEST_REQ, 1, 4, rpm_test_xlate, rpm_test_apply, 0);Swevent.c (x:\j3_ctc\nhlos\rpm_proc\core\power\rpm\common): rpm_register_resource(RPM_SWEVENT_REQ, 1, sizeof(swevent_inrep), rpm_swevent_xlate, rpm_trace_control, 0);
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
所有的资源都是通过rpm_register_resource函数注册。可以看到像RPM_BUS_SPDM_CLK_REQ,RPM_GPIO_TOGGLE_REQ都是通过这个接口直接注册的。
- NPA相关也是通过rpm_register_resource注册,但是用了NPA的driver去实际注册资源和使用资源。
NPA相关的两个注册接口是:rpm_create_npa_adapter(),rpm_create_npa_settling_adapter()。
可以看到这两个就注册了三个CLOCK相关的资源,
---- rpm_create_npa_adapter Matches (8 in 5 files) ----ClockRPM.c : clk0_adapter = rpm_create_npa_adapter(RPM_CLOCK_0_REQ, 3)ClockRPM.c : clk1_adapter = rpm_create_npa_adapter(RPM_CLOCK_1_REQ, 2)Rpmserver.h :rpm_npa_adapter rpm_create_npa_adapter(rpm_resource_type resource, unsigned num_npa_resources)Rpm_npa.cpp :rpm_npa_adapter rpm_create_npa_adapter(rpm_resource_type resource, unsigned num_npa_resources)---- rpm_create_npa_settling_adapter Matches (5 in 5 files) ----ClockRPM.c : clk2_adapter = rpm_create_npa_settling_adapter(RPM_CLOCK_2_REQ, 1)Rpmserver.h :rpm_npa_adapter rpm_create_npa_settling_adapter(rpm_resource_type resource, unsigned num_npa_resources)Rpm_npa.cpp :rpm_npa_adapter rpm_create_npa_settling_adapter(rpm_resource_type resource, unsigned num_npa_resources)
NPA client的创建函数是:npa_create_sync_client()
NPA clien request的函数是:npa_issue_required_request()
NPA client创建并request这个资源,必须要适用像下面这样的npa node。这个可以直接像下面这样定义。然后再使用。步骤如下:
1.定义npa node definition
static npa_resource_definition sleep_uber_resource[] = { { "/sleep/uber", "on/off", 0x7, &npa_or_plugin, NPA_RESOURCE_DEFAULT, NULL, } }; npa_node_definition sleep_uber_node = { "/node/sleep/uber", sleep_uber_driver, NPA_NODE_DEFAULT, NULL, 0, NULL, NPA_ARRAY(sleep_uber_resource) };
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
2.npa_define_node(&sleep_uber_node, initial_state, NULL),调用这个函数初始化这个NPA node
3.创建Client
uber_node_handle = npa_create_sync_client("/sleep/uber", "sleep", NPA_CLIENT_REQUIRED)
4.npa_issue_required_request(uber_node_handle, request) : request
RPM Resource handle
当子系统通过share memory发送请求给RPM。RPM负责处理这些请求并设置。RPM处理请求的流程如下:
SMD IRQ ->smd_isr()->rpm_smd_handler() [在Handler::init函数中注册的,rpm_request channel的SMD处理函数]->Handler::queue()->schedule_me()->Handler::execute_until()->Handler::processMessage()->resource_ee_request()->发到每个资源注册xlate然后再调用apply等
下面以高通控制DDR频率(BIMC)的过程为例,看一下kernel这边怎么发送请求给RPM的。
device tree设置如下:
cpubw: qcom,cpubw@0 { reg = <0 4>; compatible = "qcom,devbw"; governor = "cpufreq"; qcom,src-dst-ports = <1 512>; qcom,active-only; qcom,bw-tbl = < 762 >, < 1525 >, < 3051 >, < 4066 >; };
对应的kernel代码在devfeq_devbw.c文件。根据算法算出当前应该设定的ddr总线频率之后,最后通过以下顺序发消息给RPM。这里暂时不讨论按什么规则选择需要的频率的,只看按什么路径发频率给RPM的。
set_bw()->msm_bus_scale_client_update_request()->update_request_adhoc()[msm_bus_arb_adhoc.c]->update_path()->msm_bus_commit_data()->flush_bw_data()->send_rpm_msg()->msm_rpm_send_message()
System Sleep Overview
RPM除了对Clock和LDO等资源的管理之外,还管理整个系统睡眠。
可以看到睡眠并不是通过shared memory发送消息给RPM的,而是子系统通过设置对应的SPM,SPM触发RPM相应的中断来完成的。
kernel这边设置spm的模式的设置如下:
enum { MSM_SPM_MODE_DISABLED, MSM_SPM_MODE_CLOCK_GATING, MSM_SPM_MODE_RETENTION, MSM_SPM_MODE_GDHS, MSM_SPM_MODE_POWER_COLLAPSE, MSM_SPM_MODE_NR};
rpm中,与哪几个子系统传递接收message,然后和哪几个子系统的SPM进行交互,交互的中断号是多少的设置如下:
static SystemData temp_config_data ={ .num_ees = 4, .ees = (EEData[] ) { [0] = { .edge = SMD_APPS_RPM, .smd_fifo_sz = 1024, .ee_buflen = 256, .priority = 4, .wakeupInt = (1 << 5) | (1 << 7), .spm = { .numCores = 1, .bringupInts = (unsigned[]) { 15 }, .bringupAcks = (unsigned[]) { 20 }, .shutdownInts = (unsigned[]) { 14 }, .shutdownAcks = (unsigned[]) { 4 }, }, }, [1] = { .edge = SMD_MODEM_RPM, .smd_fifo_sz = 1024, .ee_buflen = 1024, .priority = 2, .wakeupInt = (1 << 13) | (1 << 15), .spm = { .numCores = 1, .bringupInts = (unsigned[]) { 25 }, .bringupAcks = (unsigned[]) { 22 }, .shutdownInts = (unsigned[]) { 24 }, .shutdownAcks = (unsigned[]) { 6 }, }, }, [2] = { .edge = SMD_RIVA_RPM, .smd_fifo_sz = 1024, .ee_buflen = 256, .priority = 1, .wakeupInt = (1 << 17) | (1 << 19), .spm = { .numCores = 1, .bringupInts = (unsigned[]) { 31 }, .bringupAcks = (unsigned[]) { 23 }, .shutdownInts = (unsigned[]) { 30 }, .shutdownAcks = (unsigned[]) { 7 }, }, }, [3] = { .edge = SMD_RPM_TZ, .smd_fifo_sz = 1024, .ee_buflen = 256, .priority = 5, .wakeupInt = 0, .spm = { .numCores = 0, .bringupInts = (unsigned[]) { 31 }, .bringupAcks = (unsigned[]) { 23 }, .shutdownInts = (unsigned[]) { 30 }, .shutdownAcks = (unsigned[]) { 7 }, }, }, }, .supported_classes = SUPPORTED_CLASSES, .supported_resources = SUPPORTED_RESOURCES, .classes = (ResourceClassData[SUPPORTED_CLASSES]) { 0 }, .resources = (ResourceData[SUPPORTED_RESOURCES]) { 0 }, .resource_seeds = (int16_t[SUPPORTED_RESOURCES]) { 0 },}
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
- 29
- 30
- 31
- 32
- 33
- 34
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- 44
- 45
- 46
- 47
- 48
- 49
- 50
- 51
- 52
- 53
- 54
- 55
- 56
- 57
- 58
- 59
- 60
- 61
- 62
- 63
- 64
- 65
- 66
- 67
但还不知道这些中断号到底怎么来的,,中断的个数都是1个,,这个怎么来的也还是不知~
这些内容在rpm_spm_init中会读取,然后设置中断。shutdownISR对应的中断处理函数为rpm_spm_shutdown_high_isr。
这个中断函数会调用rpm_spm_state_machine()处理RPM状态机,进入或者阻止进入睡眠模式等。
void rpm_spm_state_machine(unsigned ee, rpm_spm_entry_reason reason){ INTLOCK(); bool changed_state = false; EEData *ee_state = &(rpm->ees[ee]); SetChanger *changer = ee_state->changer; do { switch(ee_state->subsystem_status) { case SPM_AWAKE: changed_state = FALSE; if(0 == ee_state->num_active_cores) { SPM_CHANGE_STATE(SPM_GOING_TO_SLEEP); } else { rpm_acknowledge_spm_handshakes(ee); } break; case SPM_GOING_TO_SLEEP: if(changed_state) { uint64_t deadline = 0; if(! rpm_get_wakeup_deadline(ee, deadline)) { deadline = 0; } changer->setWakeTime (deadline); changer->enqueue(RPM_SLEEP_SET, 0); } changed_state = FALSE; if((SPM_TRANSITION_COMPLETE == reason) && (RPM_SLEEP_SET == changer->currentSet())) { SPM_CHANGE_STATE(SPM_SLEEPING); } if(SPM_BRINGUP_REQ == reason) { theSchedule().preempt(); } break; case SPM_SLEEPING: if(changed_state) { uint64_t deadline = changer->getWakeTime (); changer->enqueue(RPM_ACTIVE_SET, deadline); } changed_state = FALSE; if(ee_state->num_active_cores > 0) { SPM_CHANGE_STATE(SPM_WAKING_UP); } break; case SPM_WAKING_UP: if(changed_state) { if(RPM_SLEEP_SET == changer->currentSet() || changer->inTransition()) { changer->enqueue(RPM_ACTIVE_SET, 0); } } changed_state = FALSE; if(RPM_ACTIVE_SET == changer->currentSet() && !changer->inTransition()) { SPM_CHANGE_STATE(SPM_AWAKE); } break; } } while(changed_state); INTFREE();}
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
- 29
- 30
- 31
- 32
- 33
- 34
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- 44
- 45
- 46
- 47
- 48
- 49
- 50
- 51
- 52
- 53
- 54
- 55
- 56
- 57
- 58
- 59
- 60
- 61
- 62
- 63
- 64
- 65
- 66
- 67
- 68
- 69
- 70
- 71
- 72
- 73
- 74
- 75
- 76
- 77
- 78
- 79
- 80
- 81
- 82
- 83
- 84
- 85
- 86
- 87
- 88
- 89
- 90
- 91
- 92
- 93
- 94
- 95
- 96
最终会跑到SetChanger::enqueue()里。
RPM与MPM的交互
RPM 状态读取
rpm的状态读取,在/d/rpm_stats里边可以读到AP这边设置的几个APSS,MPSS,PRONTO等几个对应的RPM状态。
读的内容当然也是从rpm_request这个shared memory里边读的。相应的设置在msm8916-pm.dts文件里边有。
qcom,rpm-stats@29dba0 { compatible = "qcom,rpm-stats"; reg = <0x29dba0 0x1000>; reg-names = "phys_addr_base"; qcom,sleep-stats-version = <2>; }; qcom,rpm-master-stats@60150 { compatible = "qcom,rpm-master-stats"; reg = <0x60150 0x2030>; qcom,masters = "APSS", "MPSS", "PRONTO"; qcom,master-stats-version = <2>; qcom,master-offset = <4096>; }; qcom,rpm-rbcpr-stats@0x29daa0 { compatible = "qcom,rpmrbcpr-stats"; reg = <0x29daa0 0x1a0000>; qcom,start-offset = <0x190010>; };
0 0