2016-11-25 09:33:51 +00:00
// Copyright 2015-2016 Espressif Systems (Shanghai) PTE LTD
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
# include "sdkconfig.h"
# include <stdint.h>
# include <stdio.h>
# include <stdlib.h>
# include <stdbool.h>
# include <string.h>
# include "freertos/FreeRTOS.h"
# include "freertos/task.h"
# include <esp_types.h>
# include "esp_err.h"
2017-03-22 03:07:37 +00:00
//#define LOG_LOCAL_LEVEL ESP_LOG_VERBOSE
2016-11-25 09:33:51 +00:00
# include "esp_log.h"
2019-03-26 08:30:43 +00:00
# include "esp_intr_alloc.h"
2016-11-25 09:33:51 +00:00
# include "esp_attr.h"
2018-09-20 04:13:43 +00:00
# include "esp_ipc.h"
2016-11-25 09:33:51 +00:00
# include <limits.h>
# include <assert.h>
2019-04-12 04:21:44 +00:00
# include "soc/soc.h"
2016-11-25 09:33:51 +00:00
static const char * TAG = " intr_alloc " ;
# define ETS_INTERNAL_TIMER0_INTR_NO 6
# define ETS_INTERNAL_TIMER1_INTR_NO 15
# define ETS_INTERNAL_TIMER2_INTR_NO 16
# define ETS_INTERNAL_SW0_INTR_NO 7
# define ETS_INTERNAL_SW1_INTR_NO 29
# define ETS_INTERNAL_PROFILING_INTR_NO 11
/*
Define this to debug the choices made when allocating the interrupt . This leads to much debugging
output within a critical region , which can lead to weird effects like e . g . the interrupt watchdog
being triggered , that is why it is separate from the normal LOG * scheme .
*/
//define DEBUG_INT_ALLOC_DECISIONS
# ifdef DEBUG_INT_ALLOC_DECISIONS
# define ALCHLOG(...) ESP_EARLY_LOGD(TAG, __VA_ARGS__)
# else
# define ALCHLOG(...) do {} while (0)
# endif
typedef enum {
INTDESC_NORMAL = 0 ,
INTDESC_RESVD ,
INTDESC_SPECIAL //for xtensa timers / software ints
} int_desc_flag_t ;
typedef enum {
INTTP_LEVEL = 0 ,
INTTP_EDGE ,
INTTP_NA
} int_type_t ;
typedef struct {
int level ;
int_type_t type ;
int_desc_flag_t cpuflags [ 2 ] ;
} int_desc_t ;
2017-03-22 03:07:37 +00:00
//We should mark the interrupt for the timer used by FreeRTOS as reserved. The specific timer
2016-11-25 09:33:51 +00:00
//is selectable using menuconfig; we use these cpp bits to convert that into something we can use in
//the table below.
# if CONFIG_FREERTOS_CORETIMER_0
# define INT6RES INTDESC_RESVD
# else
# define INT6RES INTDESC_SPECIAL
# endif
# if CONFIG_FREERTOS_CORETIMER_1
# define INT15RES INTDESC_RESVD
# else
# define INT15RES INTDESC_SPECIAL
# endif
//This is basically a software-readable version of the interrupt usage table in include/soc/soc.h
const static int_desc_t int_desc [ 32 ] = {
{ 1 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //0
{ 1 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //1
2016-12-08 04:38:22 +00:00
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //2
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //3
2016-11-25 09:33:51 +00:00
{ 1 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_NORMAL } } , //4
2017-06-15 09:20:25 +00:00
{ 1 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //5
2016-11-25 09:33:51 +00:00
{ 1 , INTTP_NA , { INT6RES , INT6RES } } , //6
{ 1 , INTTP_NA , { INTDESC_SPECIAL , INTDESC_SPECIAL } } , //7
{ 1 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //8
2017-08-21 14:29:50 +00:00
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //9
{ 1 , INTTP_EDGE , { INTDESC_NORMAL , INTDESC_NORMAL } } , //10
2016-11-25 09:33:51 +00:00
{ 3 , INTTP_NA , { INTDESC_SPECIAL , INTDESC_SPECIAL } } , //11
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //12
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //13
{ 7 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //14, NMI
{ 3 , INTTP_NA , { INT15RES , INT15RES } } , //15
2017-01-15 03:11:58 +00:00
{ 5 , INTTP_NA , { INTDESC_SPECIAL , INTDESC_SPECIAL } } , //16
2016-11-25 09:33:51 +00:00
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //17
{ 1 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //18
{ 2 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //19
{ 2 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //20
{ 2 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //21
{ 3 , INTTP_EDGE , { INTDESC_RESVD , INTDESC_NORMAL } } , //22
{ 3 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_NORMAL } } , //23
{ 4 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_NORMAL } } , //24
{ 4 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //25
2019-02-09 11:21:41 +00:00
{ 5 , INTTP_LEVEL , { INTDESC_NORMAL , INTDESC_RESVD } } , //26
2016-11-25 09:33:51 +00:00
{ 3 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //27
{ 4 , INTTP_EDGE , { INTDESC_NORMAL , INTDESC_NORMAL } } , //28
{ 3 , INTTP_NA , { INTDESC_SPECIAL , INTDESC_SPECIAL } } , //29
{ 4 , INTTP_EDGE , { INTDESC_RESVD , INTDESC_RESVD } } , //30
{ 5 , INTTP_LEVEL , { INTDESC_RESVD , INTDESC_RESVD } } , //31
} ;
typedef struct shared_vector_desc_t shared_vector_desc_t ;
typedef struct vector_desc_t vector_desc_t ;
struct shared_vector_desc_t {
2016-12-07 13:30:21 +00:00
int disabled : 1 ;
int source : 8 ;
2016-11-25 09:33:51 +00:00
volatile uint32_t * statusreg ;
uint32_t statusmask ;
intr_handler_t isr ;
void * arg ;
shared_vector_desc_t * next ;
} ;
# define VECDESC_FL_RESERVED (1<<0)
# define VECDESC_FL_INIRAM (1<<1)
# define VECDESC_FL_SHARED (1<<2)
# define VECDESC_FL_NONSHARED (1<<3)
2016-12-07 13:30:21 +00:00
//Pack using bitfields for better memory use
2016-11-25 09:33:51 +00:00
struct vector_desc_t {
2016-12-07 13:30:21 +00:00
int flags : 16 ; //OR of VECDESC_FLAG_* defines
unsigned int cpu : 1 ;
unsigned int intno : 5 ;
int source : 8 ; //Interrupt mux flags, used when not shared
2016-11-25 09:33:51 +00:00
shared_vector_desc_t * shared_vec_info ; //used when VECDESC_FL_SHARED
vector_desc_t * next ;
} ;
2016-12-07 13:30:21 +00:00
struct intr_handle_data_t {
2016-11-25 09:33:51 +00:00
vector_desc_t * vector_desc ;
shared_vector_desc_t * shared_vector_desc ;
} ;
2017-03-22 03:07:37 +00:00
typedef struct non_shared_isr_arg_t non_shared_isr_arg_t ;
struct non_shared_isr_arg_t {
intr_handler_t isr ;
void * isr_arg ;
int source ;
} ;
2016-11-25 09:33:51 +00:00
2016-12-07 13:30:21 +00:00
//Linked list of vector descriptions, sorted by cpu.intno value
2017-08-18 07:15:47 +00:00
static vector_desc_t * vector_desc_head = NULL ;
2016-11-25 09:33:51 +00:00
//This bitmask has an 1 if the int should be disabled when the flash is disabled.
static uint32_t non_iram_int_mask [ portNUM_PROCESSORS ] ;
//This bitmask has 1 in it if the int was disabled using esp_intr_noniram_disable.
static uint32_t non_iram_int_disabled [ portNUM_PROCESSORS ] ;
static bool non_iram_int_disabled_flag [ portNUM_PROCESSORS ] ;
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
extern uint32_t port_switch_flag [ ] ;
# endif
2016-11-25 09:33:51 +00:00
static portMUX_TYPE spinlock = portMUX_INITIALIZER_UNLOCKED ;
//Inserts an item into vector_desc list so that the list is sorted
2016-12-07 13:30:21 +00:00
//with an incrementing cpu.intno value.
2017-03-22 03:07:37 +00:00
static void insert_vector_desc ( vector_desc_t * to_insert )
2016-11-25 09:33:51 +00:00
{
vector_desc_t * vd = vector_desc_head ;
vector_desc_t * prev = NULL ;
while ( vd ! = NULL ) {
2016-12-07 13:30:21 +00:00
if ( vd - > cpu > to_insert - > cpu ) break ;
if ( vd - > cpu = = to_insert - > cpu & & vd - > intno > = to_insert - > intno ) break ;
2016-11-25 09:33:51 +00:00
prev = vd ;
vd = vd - > next ;
}
2018-05-13 03:46:09 +00:00
if ( ( vector_desc_head = = NULL ) | | ( prev = = NULL ) ) {
2016-11-25 09:33:51 +00:00
//First item
2018-05-13 03:46:09 +00:00
to_insert - > next = vd ;
2016-11-25 09:33:51 +00:00
vector_desc_head = to_insert ;
} else {
prev - > next = to_insert ;
to_insert - > next = vd ;
}
}
//Returns a vector_desc entry for an intno/cpu, or NULL if none exists.
2017-03-22 03:07:37 +00:00
static vector_desc_t * find_desc_for_int ( int intno , int cpu )
2016-11-25 09:33:51 +00:00
{
vector_desc_t * vd = vector_desc_head ;
while ( vd ! = NULL ) {
2016-12-07 13:30:21 +00:00
if ( vd - > cpu = = cpu & & vd - > intno = = intno ) break ;
2016-11-25 09:33:51 +00:00
vd = vd - > next ;
}
return vd ;
}
//Returns a vector_desc entry for an intno/cpu.
//Either returns a preexisting one or allocates a new one and inserts
2016-12-06 06:20:12 +00:00
//it into the list. Returns NULL on malloc fail.
2017-03-22 03:07:37 +00:00
static vector_desc_t * get_desc_for_int ( int intno , int cpu )
2016-11-25 09:33:51 +00:00
{
vector_desc_t * vd = find_desc_for_int ( intno , cpu ) ;
if ( vd = = NULL ) {
2017-09-22 08:02:39 +00:00
vector_desc_t * newvd = heap_caps_malloc ( sizeof ( vector_desc_t ) , MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT ) ;
2016-12-06 06:20:12 +00:00
if ( newvd = = NULL ) return NULL ;
2016-11-25 09:33:51 +00:00
memset ( newvd , 0 , sizeof ( vector_desc_t ) ) ;
2016-12-07 13:30:21 +00:00
newvd - > intno = intno ;
newvd - > cpu = cpu ;
2016-11-25 09:33:51 +00:00
insert_vector_desc ( newvd ) ;
return newvd ;
} else {
return vd ;
}
}
2017-08-18 07:15:47 +00:00
//Returns a vector_desc entry for an source, the cpu parameter is used to tell GPIO_INT and GPIO_NMI from different CPUs
static vector_desc_t * find_desc_for_source ( int source , int cpu )
{
vector_desc_t * vd = vector_desc_head ;
while ( vd ! = NULL ) {
if ( ! ( vd - > flags & VECDESC_FL_SHARED ) ) {
if ( vd - > source = = source & & cpu = = vd - > cpu ) break ;
} else if ( vd - > cpu = = cpu ) {
// check only shared vds for the correct cpu, otherwise skip
bool found = false ;
shared_vector_desc_t * svd = vd - > shared_vec_info ;
assert ( svd ! = NULL ) ;
while ( svd ) {
if ( svd - > source = = source ) {
found = true ;
break ;
}
svd = svd - > next ;
}
if ( found ) break ;
}
vd = vd - > next ;
}
return vd ;
}
2016-11-25 09:33:51 +00:00
esp_err_t esp_intr_mark_shared ( int intno , int cpu , bool is_int_ram )
{
if ( intno > 31 ) return ESP_ERR_INVALID_ARG ;
if ( cpu > = portNUM_PROCESSORS ) return ESP_ERR_INVALID_ARG ;
portENTER_CRITICAL ( & spinlock ) ;
vector_desc_t * vd = get_desc_for_int ( intno , cpu ) ;
2016-12-06 06:20:12 +00:00
if ( vd = = NULL ) {
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_ERR_NO_MEM ;
2017-03-22 03:07:37 +00:00
}
2016-11-25 09:33:51 +00:00
vd - > flags = VECDESC_FL_SHARED ;
if ( is_int_ram ) vd - > flags | = VECDESC_FL_INIRAM ;
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_OK ;
}
esp_err_t esp_intr_reserve ( int intno , int cpu )
{
if ( intno > 31 ) return ESP_ERR_INVALID_ARG ;
if ( cpu > = portNUM_PROCESSORS ) return ESP_ERR_INVALID_ARG ;
portENTER_CRITICAL ( & spinlock ) ;
vector_desc_t * vd = get_desc_for_int ( intno , cpu ) ;
2016-12-06 06:20:12 +00:00
if ( vd = = NULL ) {
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_ERR_NO_MEM ;
2017-03-22 03:07:37 +00:00
}
2016-11-25 09:33:51 +00:00
vd - > flags = VECDESC_FL_RESERVED ;
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_OK ;
}
2017-03-22 03:07:37 +00:00
//Interrupt handler table and unhandled uinterrupt routine. Duplicated
//from xtensa_intr.c... it's supposed to be private, but we need to look
//into it in order to see if someone allocated an int using
2016-11-25 09:33:51 +00:00
//xt_set_interrupt_handler.
typedef struct xt_handler_table_entry {
void * handler ;
void * arg ;
} xt_handler_table_entry ;
extern xt_handler_table_entry _xt_interrupt_table [ XCHAL_NUM_INTERRUPTS * portNUM_PROCESSORS ] ;
extern void xt_unhandled_interrupt ( void * arg ) ;
//Returns true if handler for interrupt is not the default unhandled interrupt handler
2017-03-22 03:07:37 +00:00
static bool int_has_handler ( int intr , int cpu )
2016-11-25 09:33:51 +00:00
{
return ( _xt_interrupt_table [ intr * portNUM_PROCESSORS + cpu ] . handler ! = xt_unhandled_interrupt ) ;
}
2017-08-18 07:15:47 +00:00
static bool is_vect_desc_usable ( vector_desc_t * vd , int flags , int cpu , int force )
{
//Check if interrupt is not reserved by design
int x = vd - > intno ;
if ( int_desc [ x ] . cpuflags [ cpu ] = = INTDESC_RESVD ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: reserved " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
if ( int_desc [ x ] . cpuflags [ cpu ] = = INTDESC_SPECIAL & & force = = - 1 ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: special-purpose int " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
//Check if the interrupt level is acceptable
if ( ! ( flags & ( 1 < < int_desc [ x ] . level ) ) ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: incompatible level " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
//check if edge/level type matches what we want
if ( ( ( flags & ESP_INTR_FLAG_EDGE ) & & ( int_desc [ x ] . type = = INTTP_LEVEL ) ) | |
( ( ( ! ( flags & ESP_INTR_FLAG_EDGE ) ) & & ( int_desc [ x ] . type = = INTTP_EDGE ) ) ) ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: incompatible trigger type " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
//check if interrupt is reserved at runtime
if ( vd - > flags & VECDESC_FL_RESERVED ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: reserved at runtime. " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
2019-04-08 10:02:05 +00:00
2017-08-18 07:15:47 +00:00
//Ints can't be both shared and non-shared.
assert ( ! ( ( vd - > flags & VECDESC_FL_SHARED ) & & ( vd - > flags & VECDESC_FL_NONSHARED ) ) ) ;
//check if interrupt already is in use by a non-shared interrupt
if ( vd - > flags & VECDESC_FL_NONSHARED ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: already in (non-shared) use. " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
// check shared interrupt flags
if ( vd - > flags & VECDESC_FL_SHARED ) {
if ( flags & ESP_INTR_FLAG_SHARED ) {
bool in_iram_flag = ( ( flags & ESP_INTR_FLAG_IRAM ) ! = 0 ) ;
bool desc_in_iram_flag = ( ( vd - > flags & VECDESC_FL_INIRAM ) ! = 0 ) ;
//Bail out if int is shared, but iram property doesn't match what we want.
if ( ( vd - > flags & VECDESC_FL_SHARED ) & & ( desc_in_iram_flag ! = in_iram_flag ) ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: shared but iram prop doesn't match " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
} else {
//We need an unshared IRQ; can't use shared ones; bail out if this is shared.
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...Unusable: int is shared, we need non-shared. " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
} else if ( int_has_handler ( x , cpu ) ) {
//Check if interrupt already is allocated by xt_set_interrupt_handler
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ....Unusable: already allocated " ) ;
2017-08-18 07:15:47 +00:00
return false ;
}
2019-04-08 10:02:05 +00:00
2017-08-18 07:15:47 +00:00
return true ;
}
2016-11-25 09:33:51 +00:00
//Locate a free interrupt compatible with the flags given.
//The 'force' argument can be -1, or 0-31 to force checking a certain interrupt.
2016-12-07 13:30:21 +00:00
//When a CPU is forced, the INTDESC_SPECIAL marked interrupts are also accepted.
2017-08-18 07:15:47 +00:00
static int get_available_int ( int flags , int cpu , int force , int source )
2016-11-25 09:33:51 +00:00
{
int x ;
int best = - 1 ;
int bestLevel = 9 ;
int bestSharedCt = INT_MAX ;
//Default vector desc, for vectors not in the linked list
vector_desc_t empty_vect_desc ;
memset ( & empty_vect_desc , 0 , sizeof ( vector_desc_t ) ) ;
2017-08-18 07:15:47 +00:00
2019-04-08 10:02:05 +00:00
2016-11-25 09:33:51 +00:00
//Level defaults to any low/med interrupt
if ( ! ( flags & ESP_INTR_FLAG_LEVELMASK ) ) flags | = ESP_INTR_FLAG_LOWMED ;
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_available_int: try to find existing. Cpu: %d, Source: %d " , cpu , source ) ;
2017-08-18 07:15:47 +00:00
vector_desc_t * vd = find_desc_for_source ( source , cpu ) ;
if ( vd ) {
// if existing vd found, don't need to search any more.
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_avalible_int: existing vd found. intno: %d " , vd - > intno ) ;
2017-08-18 07:15:47 +00:00
if ( force ! = - 1 & & force ! = vd - > intno ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_avalible_int: intr forced but not matach existing. existing intno: %d, force: %d " , vd - > intno , force ) ;
2017-08-18 07:15:47 +00:00
} else if ( ! is_vect_desc_usable ( vd , flags , cpu , force ) ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_avalible_int: existing vd invalid. " ) ;
2017-08-18 07:15:47 +00:00
} else {
best = vd - > intno ;
}
return best ;
}
if ( force ! = - 1 ) {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_available_int: try to find force. Cpu: %d, Source: %d, Force: %d " , cpu , source , force ) ;
2017-08-18 07:15:47 +00:00
//if force assigned, don't need to search any more.
vd = find_desc_for_int ( force , cpu ) ;
if ( vd = = NULL ) {
//if existing vd not found, just check the default state for the intr.
empty_vect_desc . intno = force ;
2019-04-08 10:02:05 +00:00
vd = & empty_vect_desc ;
2017-08-18 07:15:47 +00:00
}
if ( is_vect_desc_usable ( vd , flags , cpu , force ) ) {
best = vd - > intno ;
} else {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_avalible_int: forced vd invalid. " ) ;
2019-04-08 10:02:05 +00:00
}
2017-08-18 07:15:47 +00:00
return best ;
}
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_free_int: start looking. Current cpu: %d " , cpu ) ;
2017-08-18 07:15:47 +00:00
//No allocated handlers as well as forced intr, iterate over the 32 possible interrupts
2016-12-06 06:20:12 +00:00
for ( x = 0 ; x < 32 ; x + + ) {
2016-11-25 09:33:51 +00:00
//Grab the vector_desc for this vector.
2017-08-18 07:15:47 +00:00
vd = find_desc_for_int ( x , cpu ) ;
if ( vd = = NULL ) {
empty_vect_desc . intno = x ;
vd = & empty_vect_desc ;
2016-11-25 09:33:51 +00:00
}
2017-08-18 07:15:47 +00:00
2017-08-24 03:46:19 +00:00
ALCHLOG ( " Int %d reserved %d level %d %s hasIsr %d " ,
2017-03-22 03:07:37 +00:00
x , int_desc [ x ] . cpuflags [ cpu ] = = INTDESC_RESVD , int_desc [ x ] . level ,
2016-11-25 09:33:51 +00:00
int_desc [ x ] . type = = INTTP_LEVEL ? " LEVEL " : " EDGE " , int_has_handler ( x , cpu ) ) ;
2019-04-08 10:02:05 +00:00
2017-08-18 07:15:47 +00:00
if ( ! is_vect_desc_usable ( vd , flags , cpu , force ) ) continue ;
2016-11-25 09:33:51 +00:00
if ( flags & ESP_INTR_FLAG_SHARED ) {
//We're allocating a shared int.
2019-04-08 10:02:05 +00:00
2016-11-25 09:33:51 +00:00
//See if int already is used as a shared interrupt.
if ( vd - > flags & VECDESC_FL_SHARED ) {
//We can use this already-marked-as-shared interrupt. Count the already attached isrs in order to see
//how useful it is.
int no = 0 ;
shared_vector_desc_t * svdesc = vd - > shared_vec_info ;
while ( svdesc ! = NULL ) {
no + + ;
svdesc = svdesc - > next ;
}
if ( no < bestSharedCt | | bestLevel > int_desc [ x ] . level ) {
//Seems like this shared vector is both okay and has the least amount of ISRs already attached to it.
best = x ;
bestSharedCt = no ;
bestLevel = int_desc [ x ] . level ;
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...int %d more usable as a shared int: has %d existing vectors " , x , no ) ;
2016-11-25 09:33:51 +00:00
} else {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...worse than int %d " , best ) ;
2016-11-25 09:33:51 +00:00
}
} else {
if ( best = = - 1 ) {
2017-03-22 03:07:37 +00:00
//We haven't found a feasible shared interrupt yet. This one is still free and usable, even if
2016-11-25 09:33:51 +00:00
//not marked as shared.
//Remember it in case we don't find any other shared interrupt that qualifies.
if ( bestLevel > int_desc [ x ] . level ) {
best = x ;
bestLevel = int_desc [ x ] . level ;
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...int %d usable as a new shared int " , x ) ;
2016-11-25 09:33:51 +00:00
}
} else {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...already have a shared int " ) ;
2016-11-25 09:33:51 +00:00
}
}
} else {
//Seems this interrupt is feasible. Select it and break out of the loop; no need to search further.
if ( bestLevel > int_desc [ x ] . level ) {
best = x ;
bestLevel = int_desc [ x ] . level ;
} else {
2017-08-24 03:46:19 +00:00
ALCHLOG ( " ...worse than int %d " , best ) ;
2016-11-25 09:33:51 +00:00
}
}
}
2017-08-24 03:46:19 +00:00
ALCHLOG ( " get_available_int: using int %d " , best ) ;
2016-11-25 09:33:51 +00:00
//Okay, by now we have looked at all potential interrupts and hopefully have selected the best one in best.
return best ;
}
//Common shared isr handler. Chain-call all ISRs.
2017-03-22 03:07:37 +00:00
static void IRAM_ATTR shared_intr_isr ( void * arg )
2016-11-25 09:33:51 +00:00
{
vector_desc_t * vd = ( vector_desc_t * ) arg ;
shared_vector_desc_t * sh_vec = vd - > shared_vec_info ;
2019-03-25 10:39:55 +00:00
portENTER_CRITICAL_ISR ( & spinlock ) ;
2016-11-25 09:33:51 +00:00
while ( sh_vec ) {
2016-12-07 13:30:21 +00:00
if ( ! sh_vec - > disabled ) {
if ( ( sh_vec - > statusreg = = NULL ) | | ( * sh_vec - > statusreg & sh_vec - > statusmask ) ) {
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
traceISR_ENTER ( sh_vec - > source + ETS_INTERNAL_INTR_SOURCE_OFF ) ;
# endif
2016-12-07 13:30:21 +00:00
sh_vec - > isr ( sh_vec - > arg ) ;
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
// check if we will return to scheduler or to interrupted task after ISR
if ( ! port_switch_flag [ xPortGetCoreID ( ) ] ) {
traceISR_EXIT ( ) ;
}
# endif
2016-12-07 13:30:21 +00:00
}
2016-11-25 09:33:51 +00:00
}
2016-12-07 13:30:21 +00:00
sh_vec = sh_vec - > next ;
2016-11-25 09:33:51 +00:00
}
2019-03-25 10:39:55 +00:00
portEXIT_CRITICAL_ISR ( & spinlock ) ;
2016-11-25 09:33:51 +00:00
}
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
//Common non-shared isr handler wrapper.
static void IRAM_ATTR non_shared_intr_isr ( void * arg )
{
non_shared_isr_arg_t * ns_isr_arg = ( non_shared_isr_arg_t * ) arg ;
2019-03-25 10:39:55 +00:00
portENTER_CRITICAL_ISR ( & spinlock ) ;
2017-03-22 03:07:37 +00:00
traceISR_ENTER ( ns_isr_arg - > source + ETS_INTERNAL_INTR_SOURCE_OFF ) ;
// FIXME: can we call ISR and check port_switch_flag after releasing spinlock?
// when CONFIG_SYSVIEW_ENABLE = 0 ISRs for non-shared IRQs are called without spinlock
ns_isr_arg - > isr ( ns_isr_arg - > isr_arg ) ;
// check if we will return to scheduler or to interrupted task after ISR
if ( ! port_switch_flag [ xPortGetCoreID ( ) ] ) {
traceISR_EXIT ( ) ;
}
2019-03-25 10:39:55 +00:00
portEXIT_CRITICAL_ISR ( & spinlock ) ;
2017-03-22 03:07:37 +00:00
}
# endif
2016-11-25 09:33:51 +00:00
//We use ESP_EARLY_LOG* here because this can be called before the scheduler is running.
2017-03-22 03:07:37 +00:00
esp_err_t esp_intr_alloc_intrstatus ( int source , int flags , uint32_t intrstatusreg , uint32_t intrstatusmask , intr_handler_t handler ,
void * arg , intr_handle_t * ret_handle )
2016-11-25 09:33:51 +00:00
{
2016-12-07 13:30:21 +00:00
intr_handle_data_t * ret = NULL ;
2016-11-25 09:33:51 +00:00
int force = - 1 ;
ESP_EARLY_LOGV ( TAG , " esp_intr_alloc_intrstatus (cpu %d): checking args " , xPortGetCoreID ( ) ) ;
//Shared interrupts should be level-triggered.
if ( ( flags & ESP_INTR_FLAG_SHARED ) & & ( flags & ESP_INTR_FLAG_EDGE ) ) return ESP_ERR_INVALID_ARG ;
//You can't set an handler / arg for a non-C-callable interrupt.
if ( ( flags & ESP_INTR_FLAG_HIGH ) & & ( handler ) ) return ESP_ERR_INVALID_ARG ;
2016-12-07 13:30:21 +00:00
//Shared ints should have handler and non-processor-local source
if ( ( flags & ESP_INTR_FLAG_SHARED ) & & ( ! handler | | source < 0 ) ) return ESP_ERR_INVALID_ARG ;
2016-11-25 09:33:51 +00:00
//Statusreg should have a mask
if ( intrstatusreg & & ! intrstatusmask ) return ESP_ERR_INVALID_ARG ;
2017-01-10 17:14:18 +00:00
//If the ISR is marked to be IRAM-resident, the handler must not be in the cached region
2019-04-08 10:02:05 +00:00
//ToDo: if we are to allow placing interrupt handlers into the 0x400c0000—0x400c2000 region,
//we need to make sure the interrupt is connected to the CPU0.
//CPU1 does not have access to the RTC fast memory through this region.
2017-01-10 17:14:18 +00:00
if ( ( flags & ESP_INTR_FLAG_IRAM ) & &
2019-04-08 10:02:05 +00:00
( ptrdiff_t ) handler > = SOC_RTC_IRAM_HIGH & &
( ptrdiff_t ) handler < SOC_RTC_DATA_LOW ) {
2017-01-10 17:14:18 +00:00
return ESP_ERR_INVALID_ARG ;
}
2016-11-25 09:33:51 +00:00
//Default to prio 1 for shared interrupts. Default to prio 1, 2 or 3 for non-shared interrupts.
if ( ( flags & ESP_INTR_FLAG_LEVELMASK ) = = 0 ) {
if ( flags & ESP_INTR_FLAG_SHARED ) {
flags | = ESP_INTR_FLAG_LEVEL1 ;
} else {
flags | = ESP_INTR_FLAG_LOWMED ;
}
}
ESP_EARLY_LOGV ( TAG , " esp_intr_alloc_intrstatus (cpu %d): Args okay. Resulting flags 0x%X " , xPortGetCoreID ( ) , flags ) ;
2017-03-22 03:07:37 +00:00
2016-11-25 09:33:51 +00:00
//Check 'special' interrupt sources. These are tied to one specific interrupt, so we
//have to force get_free_int to only look at that.
if ( source = = ETS_INTERNAL_TIMER0_INTR_SOURCE ) force = ETS_INTERNAL_TIMER0_INTR_NO ;
if ( source = = ETS_INTERNAL_TIMER1_INTR_SOURCE ) force = ETS_INTERNAL_TIMER1_INTR_NO ;
if ( source = = ETS_INTERNAL_TIMER2_INTR_SOURCE ) force = ETS_INTERNAL_TIMER2_INTR_NO ;
if ( source = = ETS_INTERNAL_SW0_INTR_SOURCE ) force = ETS_INTERNAL_SW0_INTR_NO ;
if ( source = = ETS_INTERNAL_SW1_INTR_SOURCE ) force = ETS_INTERNAL_SW1_INTR_NO ;
if ( source = = ETS_INTERNAL_PROFILING_INTR_SOURCE ) force = ETS_INTERNAL_PROFILING_INTR_NO ;
2016-12-07 13:30:21 +00:00
//Allocate a return handle. If we end up not needing it, we'll free it later on.
2017-09-22 08:02:39 +00:00
ret = heap_caps_malloc ( sizeof ( intr_handle_data_t ) , MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT ) ;
2016-12-07 13:30:21 +00:00
if ( ret = = NULL ) return ESP_ERR_NO_MEM ;
2016-12-06 06:20:12 +00:00
2016-11-25 09:33:51 +00:00
portENTER_CRITICAL ( & spinlock ) ;
int cpu = xPortGetCoreID ( ) ;
//See if we can find an interrupt that matches the flags.
2017-08-18 07:15:47 +00:00
int intr = get_available_int ( flags , cpu , force , source ) ;
2016-11-25 09:33:51 +00:00
if ( intr = = - 1 ) {
//None found. Bail out.
portEXIT_CRITICAL ( & spinlock ) ;
2016-12-06 06:20:12 +00:00
free ( ret ) ;
2016-11-25 09:33:51 +00:00
return ESP_ERR_NOT_FOUND ;
}
//Get an int vector desc for int.
vector_desc_t * vd = get_desc_for_int ( intr , cpu ) ;
2016-12-06 06:20:12 +00:00
if ( vd = = NULL ) {
portEXIT_CRITICAL ( & spinlock ) ;
free ( ret ) ;
return ESP_ERR_NO_MEM ;
}
2016-11-25 09:33:51 +00:00
//Allocate that int!
if ( flags & ESP_INTR_FLAG_SHARED ) {
//Populate vector entry and add to linked list.
shared_vector_desc_t * sh_vec = malloc ( sizeof ( shared_vector_desc_t ) ) ;
2016-12-06 06:20:12 +00:00
if ( sh_vec = = NULL ) {
portEXIT_CRITICAL ( & spinlock ) ;
free ( ret ) ;
return ESP_ERR_NO_MEM ;
}
2016-11-25 09:33:51 +00:00
memset ( sh_vec , 0 , sizeof ( shared_vector_desc_t ) ) ;
sh_vec - > statusreg = ( uint32_t * ) intrstatusreg ;
sh_vec - > statusmask = intrstatusmask ;
sh_vec - > isr = handler ;
sh_vec - > arg = arg ;
sh_vec - > next = vd - > shared_vec_info ;
2016-12-07 13:30:21 +00:00
sh_vec - > source = source ;
sh_vec - > disabled = 0 ;
2016-11-25 09:33:51 +00:00
vd - > shared_vec_info = sh_vec ;
vd - > flags | = VECDESC_FL_SHARED ;
//(Re-)set shared isr handler to new value.
xt_set_interrupt_handler ( intr , shared_intr_isr , vd ) ;
} else {
//Mark as unusable for other interrupt sources. This is ours now!
vd - > flags = VECDESC_FL_NONSHARED ;
if ( handler ) {
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
non_shared_isr_arg_t * ns_isr_arg = malloc ( sizeof ( non_shared_isr_arg_t ) ) ;
if ( ! ns_isr_arg ) {
portEXIT_CRITICAL ( & spinlock ) ;
free ( ret ) ;
return ESP_ERR_NO_MEM ;
}
ns_isr_arg - > isr = handler ;
ns_isr_arg - > isr_arg = arg ;
ns_isr_arg - > source = source ;
xt_set_interrupt_handler ( intr , non_shared_intr_isr , ns_isr_arg ) ;
# else
2016-11-25 09:33:51 +00:00
xt_set_interrupt_handler ( intr , handler , arg ) ;
2017-03-22 03:07:37 +00:00
# endif
2016-11-25 09:33:51 +00:00
}
if ( flags & ESP_INTR_FLAG_EDGE ) xthal_set_intclear ( 1 < < intr ) ;
2016-12-07 13:30:21 +00:00
vd - > source = source ;
2016-11-25 09:33:51 +00:00
}
if ( flags & ESP_INTR_FLAG_IRAM ) {
vd - > flags | = VECDESC_FL_INIRAM ;
non_iram_int_mask [ cpu ] & = ~ ( 1 < < intr ) ;
} else {
vd - > flags & = ~ VECDESC_FL_INIRAM ;
non_iram_int_mask [ cpu ] | = ( 1 < < intr ) ;
}
if ( source > = 0 ) {
intr_matrix_set ( cpu , source , intr ) ;
}
2016-12-07 13:30:21 +00:00
//Fill return handle data.
ret - > vector_desc = vd ;
ret - > shared_vector_desc = vd - > shared_vec_info ;
//Enable int at CPU-level;
ESP_INTR_ENABLE ( intr ) ;
//If interrupt has to be started disabled, do that now; ints won't be enabled for real until the end
//of the critical section.
if ( flags & ESP_INTR_FLAG_INTRDISABLED ) {
esp_intr_disable ( ret ) ;
}
portEXIT_CRITICAL ( & spinlock ) ;
//Fill return handle if needed, otherwise free handle.
2016-11-25 09:33:51 +00:00
if ( ret_handle ! = NULL ) {
* ret_handle = ret ;
2016-12-07 13:30:21 +00:00
} else {
free ( ret ) ;
2016-11-25 09:33:51 +00:00
}
ESP_EARLY_LOGD ( TAG , " Connected src %d to int %d (cpu %d) " , source , intr , cpu ) ;
return ESP_OK ;
}
2017-03-22 03:07:37 +00:00
esp_err_t esp_intr_alloc ( int source , int flags , intr_handler_t handler , void * arg , intr_handle_t * ret_handle )
2016-11-25 09:33:51 +00:00
{
/*
As an optimization , we can create a table with the possible interrupt status registers and masks for every single
2017-03-22 03:07:37 +00:00
source there is . We can then add code here to look up an applicable value and pass that to the
2016-11-25 09:33:51 +00:00
esp_intr_alloc_intrstatus function .
*/
return esp_intr_alloc_intrstatus ( source , flags , 0 , 0 , handler , arg , ret_handle ) ;
}
2017-11-09 11:21:39 +00:00
esp_err_t IRAM_ATTR esp_intr_set_in_iram ( intr_handle_t handle , bool is_in_iram )
{
if ( ! handle ) return ESP_ERR_INVALID_ARG ;
vector_desc_t * vd = handle - > vector_desc ;
if ( vd - > flags & VECDESC_FL_SHARED ) {
return ESP_ERR_INVALID_ARG ;
}
portENTER_CRITICAL ( & spinlock ) ;
uint32_t mask = ( 1 < < vd - > intno ) ;
if ( is_in_iram ) {
vd - > flags | = VECDESC_FL_INIRAM ;
non_iram_int_mask [ vd - > cpu ] & = ~ mask ;
} else {
vd - > flags & = ~ VECDESC_FL_INIRAM ;
non_iram_int_mask [ vd - > cpu ] | = mask ;
}
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_OK ;
}
2016-11-25 09:33:51 +00:00
2018-10-13 06:38:48 +00:00
static void esp_intr_free_cb ( void * arg )
{
( void ) esp_intr_free ( ( intr_handle_t ) arg ) ;
}
2017-03-22 03:07:37 +00:00
esp_err_t esp_intr_free ( intr_handle_t handle )
2016-11-25 09:33:51 +00:00
{
bool free_shared_vector = false ;
if ( ! handle ) return ESP_ERR_INVALID_ARG ;
2018-09-20 04:13:43 +00:00
//Assign this routine to the core where this interrupt is allocated on.
if ( handle - > vector_desc - > cpu ! = xPortGetCoreID ( ) ) {
2018-10-13 06:38:48 +00:00
esp_err_t ret = esp_ipc_call_blocking ( handle - > vector_desc - > cpu , & esp_intr_free_cb , ( void * ) handle ) ;
2018-09-20 04:13:43 +00:00
return ret = = ESP_OK ? ESP_OK : ESP_FAIL ;
}
2016-11-25 09:33:51 +00:00
portENTER_CRITICAL ( & spinlock ) ;
2016-12-08 04:04:26 +00:00
esp_intr_disable ( handle ) ;
2016-11-25 09:33:51 +00:00
if ( handle - > vector_desc - > flags & VECDESC_FL_SHARED ) {
2017-03-22 03:07:37 +00:00
//Find and kill the shared int
2016-11-25 09:33:51 +00:00
shared_vector_desc_t * svd = handle - > vector_desc - > shared_vec_info ;
shared_vector_desc_t * prevsvd = NULL ;
assert ( svd ) ; //should be something in there for a shared int
while ( svd ! = NULL ) {
if ( svd = = handle - > shared_vector_desc ) {
//Found it. Now kill it.
if ( prevsvd ) {
prevsvd - > next = svd - > next ;
} else {
handle - > vector_desc - > shared_vec_info = svd - > next ;
}
free ( svd ) ;
break ;
}
prevsvd = svd ;
svd = svd - > next ;
}
//If nothing left, disable interrupt.
if ( handle - > vector_desc - > shared_vec_info = = NULL ) free_shared_vector = true ;
ESP_LOGV ( TAG , " esp_intr_free: Deleting shared int: %s. Shared int is %s " , svd ? " not found or last one " : " deleted " , free_shared_vector ? " empty now. " : " still in use " ) ;
}
if ( ( handle - > vector_desc - > flags & VECDESC_FL_NONSHARED ) | | free_shared_vector ) {
ESP_LOGV ( TAG , " esp_intr_free: Disabling int, killing handler " ) ;
2017-03-22 03:07:37 +00:00
# if CONFIG_SYSVIEW_ENABLE
if ( ! free_shared_vector ) {
void * isr_arg = xt_get_interrupt_handler_arg ( handle - > vector_desc - > intno ) ;
if ( isr_arg ) {
free ( isr_arg ) ;
}
}
# endif
2016-12-07 13:30:21 +00:00
//Reset to normal handler
2016-12-08 04:38:22 +00:00
xt_set_interrupt_handler ( handle - > vector_desc - > intno , xt_unhandled_interrupt , ( void * ) ( ( int ) handle - > vector_desc - > intno ) ) ;
2016-11-25 09:33:51 +00:00
//Theoretically, we could free the vector_desc... not sure if that's worth the few bytes of memory
2017-03-22 03:07:37 +00:00
//we save.(We can also not use the same exit path for empty shared ints anymore if we delete
//the desc.) For now, just mark it as free.
2016-11-25 09:33:51 +00:00
handle - > vector_desc - > flags & = ! ( VECDESC_FL_NONSHARED | VECDESC_FL_RESERVED ) ;
//Also kill non_iram mask bit.
2016-12-07 13:30:21 +00:00
non_iram_int_mask [ handle - > vector_desc - > cpu ] & = ~ ( 1 < < ( handle - > vector_desc - > intno ) ) ;
2016-11-25 09:33:51 +00:00
}
portEXIT_CRITICAL ( & spinlock ) ;
free ( handle ) ;
return ESP_OK ;
}
2016-12-07 13:30:21 +00:00
int esp_intr_get_intno ( intr_handle_t handle )
2016-11-25 09:33:51 +00:00
{
2016-12-07 13:30:21 +00:00
return handle - > vector_desc - > intno ;
2016-11-25 09:33:51 +00:00
}
2016-12-07 13:30:21 +00:00
int esp_intr_get_cpu ( intr_handle_t handle )
2016-11-25 09:33:51 +00:00
{
2016-12-07 13:30:21 +00:00
return handle - > vector_desc - > cpu ;
2016-11-25 09:33:51 +00:00
}
2016-12-07 13:30:21 +00:00
/*
Interrupt disabling strategy :
If the source is > = 0 ( meaning a muxed interrupt ) , we disable it by muxing the interrupt to a non - connected
interrupt . If the source is < 0 ( meaning an internal , per - cpu interrupt ) , we disable it using ESP_INTR_DISABLE .
This allows us to , for the muxed CPUs , disable an int from the other core . It also allows disabling shared
interrupts .
*/
//Muxing an interrupt source to interrupt 6, 7, 11, 15, 16 or 29 cause the interrupt to effectively be disabled.
# define INT_MUX_DISABLED_INTNO 6
2016-12-19 06:28:28 +00:00
esp_err_t IRAM_ATTR esp_intr_enable ( intr_handle_t handle )
2016-11-25 09:33:51 +00:00
{
if ( ! handle ) return ESP_ERR_INVALID_ARG ;
2019-03-25 10:39:55 +00:00
portENTER_CRITICAL_SAFE ( & spinlock ) ;
2016-12-07 13:30:21 +00:00
int source ;
if ( handle - > shared_vector_desc ) {
handle - > shared_vector_desc - > disabled = 0 ;
source = handle - > shared_vector_desc - > source ;
} else {
source = handle - > vector_desc - > source ;
}
if ( source > = 0 ) {
//Disabled using int matrix; re-connect to enable
intr_matrix_set ( handle - > vector_desc - > cpu , source , handle - > vector_desc - > intno ) ;
} else {
//Re-enable using cpu int ena reg
if ( handle - > vector_desc - > cpu ! = xPortGetCoreID ( ) ) return ESP_ERR_INVALID_ARG ; //Can only enable these ints on this cpu
ESP_INTR_ENABLE ( handle - > vector_desc - > intno ) ;
}
2019-03-25 10:39:55 +00:00
portEXIT_CRITICAL_SAFE ( & spinlock ) ;
2016-11-25 09:33:51 +00:00
return ESP_OK ;
}
2016-12-19 06:28:28 +00:00
esp_err_t IRAM_ATTR esp_intr_disable ( intr_handle_t handle )
2016-11-25 09:33:51 +00:00
{
if ( ! handle ) return ESP_ERR_INVALID_ARG ;
2019-03-25 10:39:55 +00:00
portENTER_CRITICAL_SAFE ( & spinlock ) ;
2016-12-07 13:30:21 +00:00
int source ;
2017-08-18 07:15:47 +00:00
bool disabled = 1 ;
2016-12-07 13:30:21 +00:00
if ( handle - > shared_vector_desc ) {
handle - > shared_vector_desc - > disabled = 1 ;
source = handle - > shared_vector_desc - > source ;
2017-08-18 07:15:47 +00:00
shared_vector_desc_t * svd = handle - > vector_desc - > shared_vec_info ;
assert ( svd ! = NULL ) ;
while ( svd ) {
if ( svd - > source = = source & & svd - > disabled = = 0 ) {
disabled = 0 ;
break ;
}
svd = svd - > next ;
}
2016-12-07 13:30:21 +00:00
} else {
source = handle - > vector_desc - > source ;
}
2017-08-18 07:15:47 +00:00
2016-12-07 13:30:21 +00:00
if ( source > = 0 ) {
2017-08-18 07:15:47 +00:00
if ( disabled ) {
//Disable using int matrix
intr_matrix_set ( handle - > vector_desc - > cpu , source , INT_MUX_DISABLED_INTNO ) ;
}
2016-12-07 13:30:21 +00:00
} else {
//Disable using per-cpu regs
if ( handle - > vector_desc - > cpu ! = xPortGetCoreID ( ) ) {
portEXIT_CRITICAL ( & spinlock ) ;
return ESP_ERR_INVALID_ARG ; //Can only enable these ints on this cpu
}
ESP_INTR_DISABLE ( handle - > vector_desc - > intno ) ;
}
2019-03-25 10:39:55 +00:00
portEXIT_CRITICAL_SAFE ( & spinlock ) ;
2016-11-25 09:33:51 +00:00
return ESP_OK ;
}
2017-01-19 17:24:55 +00:00
void IRAM_ATTR esp_intr_noniram_disable ( )
2016-11-25 09:33:51 +00:00
{
int oldint ;
int cpu = xPortGetCoreID ( ) ;
int intmask = ~ non_iram_int_mask [ cpu ] ;
2017-01-09 04:04:21 +00:00
if ( non_iram_int_disabled_flag [ cpu ] ) abort ( ) ;
2016-11-25 09:33:51 +00:00
non_iram_int_disabled_flag [ cpu ] = true ;
asm volatile (
" movi %0,0 \n "
" xsr %0,INTENABLE \n " //disable all ints first
" rsync \n "
" and a3,%0,%1 \n " //mask ints that need disabling
" wsr a3,INTENABLE \n " //write back
" rsync \n "
2017-01-18 10:31:06 +00:00
: " =&r " ( oldint ) : " r " ( intmask ) : " a3 " ) ;
2016-11-25 09:33:51 +00:00
//Save which ints we did disable
non_iram_int_disabled [ cpu ] = oldint & non_iram_int_mask [ cpu ] ;
}
2017-01-19 17:24:55 +00:00
void IRAM_ATTR esp_intr_noniram_enable ( )
2016-11-25 09:33:51 +00:00
{
int cpu = xPortGetCoreID ( ) ;
int intmask = non_iram_int_disabled [ cpu ] ;
2017-01-09 04:04:21 +00:00
if ( ! non_iram_int_disabled_flag [ cpu ] ) abort ( ) ;
2016-11-25 09:33:51 +00:00
non_iram_int_disabled_flag [ cpu ] = false ;
asm volatile (
" movi a3,0 \n "
" xsr a3,INTENABLE \n "
" rsync \n "
" or a3,a3,%0 \n "
" wsr a3,INTENABLE \n "
" rsync \n "
: : " r " ( intmask ) : " a3 " ) ;
}
2016-12-12 12:05:58 +00:00
//These functions are provided in ROM, but the ROM-based functions use non-multicore-capable
//virtualized interrupt levels. Thus, we disable them in the ld file and provide working
//equivalents here.
2016-11-25 09:33:51 +00:00
2016-12-13 03:29:34 +00:00
void IRAM_ATTR ets_isr_unmask ( unsigned int mask ) {
xt_ints_on ( mask ) ;
2016-12-12 12:05:58 +00:00
}
2016-12-13 03:29:34 +00:00
void IRAM_ATTR ets_isr_mask ( unsigned int mask ) {
xt_ints_off ( mask ) ;
2016-12-12 12:05:58 +00:00
}
2016-11-25 09:33:51 +00:00