Lint the script folder files (#5991)

This commit is contained in:
Jesse Hills 2023-12-22 20:03:47 +13:00 committed by GitHub
parent 676ae6b26e
commit d2d0058386
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
6 changed files with 398 additions and 377 deletions

View file

@ -17,28 +17,22 @@ then run this script with python3 and the files
will be generated, they still need to be formatted will be generated, they still need to be formatted
""" """
import re
import os import os
import re
import sys
from abc import ABC, abstractmethod
from pathlib import Path from pathlib import Path
from textwrap import dedent
from subprocess import call from subprocess import call
from textwrap import dedent
# Generate with # Generate with
# protoc --python_out=script/api_protobuf -I esphome/components/api/ api_options.proto # protoc --python_out=script/api_protobuf -I esphome/components/api/ api_options.proto
import aioesphomeapi.api_options_pb2 as pb import aioesphomeapi.api_options_pb2 as pb
import google.protobuf.descriptor_pb2 as descriptor import google.protobuf.descriptor_pb2 as descriptor
file_header = "// This file was automatically generated with a tool.\n" FILE_HEADER = """// This file was automatically generated with a tool.
file_header += "// See scripts/api_protobuf/api_protobuf.py\n" // See scripts/api_protobuf/api_protobuf.py
"""
cwd = Path(__file__).resolve().parent
root = cwd.parent.parent / "esphome" / "components" / "api"
prot = root / "api.protoc"
call(["protoc", "-o", str(prot), "-I", str(root), "api.proto"])
content = prot.read_bytes()
d = descriptor.FileDescriptorSet.FromString(content)
def indent_list(text, padding=" "): def indent_list(text, padding=" "):
@ -64,7 +58,7 @@ def camel_to_snake(name):
return re.sub("([a-z0-9])([A-Z])", r"\1_\2", s1).lower() return re.sub("([a-z0-9])([A-Z])", r"\1_\2", s1).lower()
class TypeInfo: class TypeInfo(ABC):
def __init__(self, field): def __init__(self, field):
self._field = field self._field = field
@ -186,10 +180,12 @@ class TypeInfo:
def dump_content(self): def dump_content(self):
o = f'out.append(" {self.name}: ");\n' o = f'out.append(" {self.name}: ");\n'
o += self.dump(f"this->{self.field_name}") + "\n" o += self.dump(f"this->{self.field_name}") + "\n"
o += f'out.append("\\n");\n' o += 'out.append("\\n");\n'
return o return o
dump = None @abstractmethod
def dump(self, name: str):
pass
TYPE_INFO = {} TYPE_INFO = {}
@ -212,7 +208,7 @@ class DoubleType(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%g", {name});\n' o = f'sprintf(buffer, "%g", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -225,7 +221,7 @@ class FloatType(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%g", {name});\n' o = f'sprintf(buffer, "%g", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -238,7 +234,7 @@ class Int64Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%lld", {name});\n' o = f'sprintf(buffer, "%lld", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -251,7 +247,7 @@ class UInt64Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%llu", {name});\n' o = f'sprintf(buffer, "%llu", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -264,7 +260,7 @@ class Int32Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%" PRId32, {name});\n' o = f'sprintf(buffer, "%" PRId32, {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -277,7 +273,7 @@ class Fixed64Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%llu", {name});\n' o = f'sprintf(buffer, "%llu", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -290,7 +286,7 @@ class Fixed32Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%" PRIu32, {name});\n' o = f'sprintf(buffer, "%" PRIu32, {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -372,7 +368,7 @@ class UInt32Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%" PRIu32, {name});\n' o = f'sprintf(buffer, "%" PRIu32, {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -406,7 +402,7 @@ class SFixed32Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%" PRId32, {name});\n' o = f'sprintf(buffer, "%" PRId32, {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -419,7 +415,7 @@ class SFixed64Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%lld", {name});\n' o = f'sprintf(buffer, "%lld", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -432,7 +428,7 @@ class SInt32Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%" PRId32, {name});\n' o = f'sprintf(buffer, "%" PRId32, {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -445,7 +441,7 @@ class SInt64Type(TypeInfo):
def dump(self, name): def dump(self, name):
o = f'sprintf(buffer, "%lld", {name});\n' o = f'sprintf(buffer, "%lld", {name});\n'
o += f"out.append(buffer);" o += "out.append(buffer);"
return o return o
@ -527,7 +523,7 @@ class RepeatedTypeInfo(TypeInfo):
def encode_content(self): def encode_content(self):
o = f"for (auto {'' if self._ti_is_bool else '&'}it : this->{self.field_name}) {{\n" o = f"for (auto {'' if self._ti_is_bool else '&'}it : this->{self.field_name}) {{\n"
o += f" buffer.{self._ti.encode_func}({self.number}, it, true);\n" o += f" buffer.{self._ti.encode_func}({self.number}, it, true);\n"
o += f"}}" o += "}"
return o return o
@property @property
@ -535,10 +531,13 @@ class RepeatedTypeInfo(TypeInfo):
o = f'for (const auto {"" if self._ti_is_bool else "&"}it : this->{self.field_name}) {{\n' o = f'for (const auto {"" if self._ti_is_bool else "&"}it : this->{self.field_name}) {{\n'
o += f' out.append(" {self.name}: ");\n' o += f' out.append(" {self.name}: ");\n'
o += indent(self._ti.dump("it")) + "\n" o += indent(self._ti.dump("it")) + "\n"
o += f' out.append("\\n");\n' o += ' out.append("\\n");\n'
o += f"}}\n" o += "}\n"
return o return o
def dump(self, _: str):
pass
def build_enum_type(desc): def build_enum_type(desc):
name = desc.name name = desc.name
@ -547,17 +546,17 @@ def build_enum_type(desc):
out += f" {v.name} = {v.number},\n" out += f" {v.name} = {v.number},\n"
out += "};\n" out += "};\n"
cpp = f"#ifdef HAS_PROTO_MESSAGE_DUMP\n" cpp = "#ifdef HAS_PROTO_MESSAGE_DUMP\n"
cpp += f"template<> const char *proto_enum_to_string<enums::{name}>(enums::{name} value) {{\n" cpp += f"template<> const char *proto_enum_to_string<enums::{name}>(enums::{name} value) {{\n"
cpp += f" switch (value) {{\n" cpp += " switch (value) {\n"
for v in desc.value: for v in desc.value:
cpp += f" case enums::{v.name}:\n" cpp += f" case enums::{v.name}:\n"
cpp += f' return "{v.name}";\n' cpp += f' return "{v.name}";\n'
cpp += f" default:\n" cpp += " default:\n"
cpp += f' return "UNKNOWN";\n' cpp += ' return "UNKNOWN";\n'
cpp += f" }}\n" cpp += " }\n"
cpp += f"}}\n" cpp += "}\n"
cpp += f"#endif\n" cpp += "#endif\n"
return out, cpp return out, cpp
@ -652,10 +651,10 @@ def build_message_type(desc):
o += f" {dump[0]} " o += f" {dump[0]} "
else: else:
o += "\n" o += "\n"
o += f" __attribute__((unused)) char buffer[64];\n" o += " __attribute__((unused)) char buffer[64];\n"
o += f' out.append("{desc.name} {{\\n");\n' o += f' out.append("{desc.name} {{\\n");\n'
o += indent("\n".join(dump)) + "\n" o += indent("\n".join(dump)) + "\n"
o += f' out.append("}}");\n' o += ' out.append("}");\n'
else: else:
o2 = f'out.append("{desc.name} {{}}");' o2 = f'out.append("{desc.name} {{}}");'
if len(o) + len(o2) + 3 < 120: if len(o) + len(o2) + 3 < 120:
@ -664,9 +663,9 @@ def build_message_type(desc):
o += "\n" o += "\n"
o += f" {o2}\n" o += f" {o2}\n"
o += "}\n" o += "}\n"
cpp += f"#ifdef HAS_PROTO_MESSAGE_DUMP\n" cpp += "#ifdef HAS_PROTO_MESSAGE_DUMP\n"
cpp += o cpp += o
cpp += f"#endif\n" cpp += "#endif\n"
prot = "#ifdef HAS_PROTO_MESSAGE_DUMP\n" prot = "#ifdef HAS_PROTO_MESSAGE_DUMP\n"
prot += "void dump_to(std::string &out) const override;\n" prot += "void dump_to(std::string &out) const override;\n"
prot += "#endif\n" prot += "#endif\n"
@ -684,71 +683,12 @@ def build_message_type(desc):
return out, cpp return out, cpp
file = d.file[0]
content = file_header
content += """\
#pragma once
#include "proto.h"
namespace esphome {
namespace api {
"""
cpp = file_header
cpp += """\
#include "api_pb2.h"
#include "esphome/core/log.h"
#include <cinttypes>
namespace esphome {
namespace api {
"""
content += "namespace enums {\n\n"
for enum in file.enum_type:
s, c = build_enum_type(enum)
content += s
cpp += c
content += "\n} // namespace enums\n\n"
mt = file.message_type
for m in mt:
s, c = build_message_type(m)
content += s
cpp += c
content += """\
} // namespace api
} // namespace esphome
"""
cpp += """\
} // namespace api
} // namespace esphome
"""
with open(root / "api_pb2.h", "w") as f:
f.write(content)
with open(root / "api_pb2.cpp", "w") as f:
f.write(cpp)
SOURCE_BOTH = 0 SOURCE_BOTH = 0
SOURCE_SERVER = 1 SOURCE_SERVER = 1
SOURCE_CLIENT = 2 SOURCE_CLIENT = 2
RECEIVE_CASES = {} RECEIVE_CASES = {}
class_name = "APIServerConnectionBase"
ifdefs = {} ifdefs = {}
@ -768,7 +708,6 @@ def build_service_message_type(mt):
ifdef = get_opt(mt, pb.ifdef) ifdef = get_opt(mt, pb.ifdef)
log = get_opt(mt, pb.log, True) log = get_opt(mt, pb.log, True)
nodelay = get_opt(mt, pb.no_delay, False)
hout = "" hout = ""
cout = "" cout = ""
@ -781,14 +720,14 @@ def build_service_message_type(mt):
# Generate send # Generate send
func = f"send_{snake}" func = f"send_{snake}"
hout += f"bool {func}(const {mt.name} &msg);\n" hout += f"bool {func}(const {mt.name} &msg);\n"
cout += f"bool {class_name}::{func}(const {mt.name} &msg) {{\n" cout += f"bool APIServerConnectionBase::{func}(const {mt.name} &msg) {{\n"
if log: if log:
cout += f"#ifdef HAS_PROTO_MESSAGE_DUMP\n" cout += "#ifdef HAS_PROTO_MESSAGE_DUMP\n"
cout += f' ESP_LOGVV(TAG, "{func}: %s", msg.dump().c_str());\n' cout += f' ESP_LOGVV(TAG, "{func}: %s", msg.dump().c_str());\n'
cout += f"#endif\n" cout += "#endif\n"
# cout += f' this->set_nodelay({str(nodelay).lower()});\n' # cout += f' this->set_nodelay({str(nodelay).lower()});\n'
cout += f" return this->send_message_<{mt.name}>(msg, {id_});\n" cout += f" return this->send_message_<{mt.name}>(msg, {id_});\n"
cout += f"}}\n" cout += "}\n"
if source in (SOURCE_BOTH, SOURCE_CLIENT): if source in (SOURCE_BOTH, SOURCE_CLIENT):
# Generate receive # Generate receive
func = f"on_{snake}" func = f"on_{snake}"
@ -797,52 +736,121 @@ def build_service_message_type(mt):
if ifdef is not None: if ifdef is not None:
case += f"#ifdef {ifdef}\n" case += f"#ifdef {ifdef}\n"
case += f"{mt.name} msg;\n" case += f"{mt.name} msg;\n"
case += f"msg.decode(msg_data, msg_size);\n" case += "msg.decode(msg_data, msg_size);\n"
if log: if log:
case += f"#ifdef HAS_PROTO_MESSAGE_DUMP\n" case += "#ifdef HAS_PROTO_MESSAGE_DUMP\n"
case += f'ESP_LOGVV(TAG, "{func}: %s", msg.dump().c_str());\n' case += f'ESP_LOGVV(TAG, "{func}: %s", msg.dump().c_str());\n'
case += f"#endif\n" case += "#endif\n"
case += f"this->{func}(msg);\n" case += f"this->{func}(msg);\n"
if ifdef is not None: if ifdef is not None:
case += f"#endif\n" case += "#endif\n"
case += "break;" case += "break;"
RECEIVE_CASES[id_] = case RECEIVE_CASES[id_] = case
if ifdef is not None: if ifdef is not None:
hout += f"#endif\n" hout += "#endif\n"
cout += f"#endif\n" cout += "#endif\n"
return hout, cout return hout, cout
hpp = file_header def main():
hpp += """\ cwd = Path(__file__).resolve().parent
#pragma once root = cwd.parent.parent / "esphome" / "components" / "api"
prot_file = root / "api.protoc"
call(["protoc", "-o", str(prot_file), "-I", str(root), "api.proto"])
proto_content = prot_file.read_bytes()
#include "api_pb2.h" # pylint: disable-next=no-member
#include "esphome/core/defines.h" d = descriptor.FileDescriptorSet.FromString(proto_content)
namespace esphome { file = d.file[0]
namespace api { content = FILE_HEADER
content += """\
#pragma once
""" #include "proto.h"
cpp = file_header namespace esphome {
cpp += """\ namespace api {
#include "api_pb2_service.h"
#include "esphome/core/log.h"
namespace esphome { """
namespace api {
static const char *const TAG = "api.service"; cpp = FILE_HEADER
cpp += """\
#include "api_pb2.h"
#include "esphome/core/log.h"
""" #include <cinttypes>
hpp += f"class {class_name} : public ProtoService {{\n" namespace esphome {
hpp += " public:\n" namespace api {
for mt in file.message_type: """
content += "namespace enums {\n\n"
for enum in file.enum_type:
s, c = build_enum_type(enum)
content += s
cpp += c
content += "\n} // namespace enums\n\n"
mt = file.message_type
for m in mt:
s, c = build_message_type(m)
content += s
cpp += c
content += """\
} // namespace api
} // namespace esphome
"""
cpp += """\
} // namespace api
} // namespace esphome
"""
with open(root / "api_pb2.h", "w", encoding="utf-8") as f:
f.write(content)
with open(root / "api_pb2.cpp", "w", encoding="utf-8") as f:
f.write(cpp)
hpp = FILE_HEADER
hpp += """\
#pragma once
#include "api_pb2.h"
#include "esphome/core/defines.h"
namespace esphome {
namespace api {
"""
cpp = FILE_HEADER
cpp += """\
#include "api_pb2_service.h"
#include "esphome/core/log.h"
namespace esphome {
namespace api {
static const char *const TAG = "api.service";
"""
class_name = "APIServerConnectionBase"
hpp += f"class {class_name} : public ProtoService {{\n"
hpp += " public:\n"
for mt in file.message_type:
obj = build_service_message_type(mt) obj = build_service_message_type(mt)
if obj is None: if obj is None:
continue continue
@ -850,35 +858,35 @@ for mt in file.message_type:
hpp += indent(hout) + "\n" hpp += indent(hout) + "\n"
cpp += cout cpp += cout
cases = list(RECEIVE_CASES.items()) cases = list(RECEIVE_CASES.items())
cases.sort() cases.sort()
hpp += " protected:\n" hpp += " protected:\n"
hpp += f" bool read_message(uint32_t msg_size, uint32_t msg_type, uint8_t *msg_data) override;\n" hpp += " bool read_message(uint32_t msg_size, uint32_t msg_type, uint8_t *msg_data) override;\n"
out = f"bool {class_name}::read_message(uint32_t msg_size, uint32_t msg_type, uint8_t *msg_data) {{\n" out = f"bool {class_name}::read_message(uint32_t msg_size, uint32_t msg_type, uint8_t *msg_data) {{\n"
out += f" switch (msg_type) {{\n" out += " switch (msg_type) {\n"
for i, case in cases: for i, case in cases:
c = f"case {i}: {{\n" c = f"case {i}: {{\n"
c += indent(case) + "\n" c += indent(case) + "\n"
c += f"}}" c += "}"
out += indent(c, " ") + "\n" out += indent(c, " ") + "\n"
out += " default:\n" out += " default:\n"
out += " return false;\n" out += " return false;\n"
out += " }\n" out += " }\n"
out += " return true;\n" out += " return true;\n"
out += "}\n" out += "}\n"
cpp += out cpp += out
hpp += "};\n" hpp += "};\n"
serv = file.service[0] serv = file.service[0]
class_name = "APIServerConnection" class_name = "APIServerConnection"
hpp += "\n" hpp += "\n"
hpp += f"class {class_name} : public {class_name}Base {{\n" hpp += f"class {class_name} : public {class_name}Base {{\n"
hpp += " public:\n" hpp += " public:\n"
hpp_protected = "" hpp_protected = ""
cpp += "\n" cpp += "\n"
m = serv.method[0] m = serv.method[0]
for m in serv.method: for m in serv.method:
func = m.name func = m.name
inp = m.input_type[1:] inp = m.input_type[1:]
ret = m.output_type[1:] ret = m.output_type[1:]
@ -916,39 +924,39 @@ for m in serv.method:
body += f"{ret} ret = this->{func}(msg);\n" body += f"{ret} ret = this->{func}(msg);\n"
ret_snake = camel_to_snake(ret) ret_snake = camel_to_snake(ret)
body += f"if (!this->send_{ret_snake}(ret)) {{\n" body += f"if (!this->send_{ret_snake}(ret)) {{\n"
body += f" this->on_fatal_error();\n" body += " this->on_fatal_error();\n"
body += "}\n" body += "}\n"
cpp += indent(body) + "\n" + "}\n" cpp += indent(body) + "\n" + "}\n"
if ifdef is not None: if ifdef is not None:
hpp += f"#endif\n" hpp += "#endif\n"
hpp_protected += f"#endif\n" hpp_protected += "#endif\n"
cpp += f"#endif\n" cpp += "#endif\n"
hpp += " protected:\n" hpp += " protected:\n"
hpp += hpp_protected hpp += hpp_protected
hpp += "};\n" hpp += "};\n"
hpp += """\ hpp += """\
} // namespace api } // namespace api
} // namespace esphome } // namespace esphome
""" """
cpp += """\ cpp += """\
} // namespace api } // namespace api
} // namespace esphome } // namespace esphome
""" """
with open(root / "api_pb2_service.h", "w") as f: with open(root / "api_pb2_service.h", "w", encoding="utf-8") as f:
f.write(hpp) f.write(hpp)
with open(root / "api_pb2_service.cpp", "w") as f: with open(root / "api_pb2_service.cpp", "w", encoding="utf-8") as f:
f.write(cpp) f.write(cpp)
prot.unlink() prot_file.unlink()
try: try:
import clang_format import clang_format
def exec_clang_format(path): def exec_clang_format(path):
@ -961,5 +969,9 @@ try:
exec_clang_format(root / "api_pb2_service.cpp") exec_clang_format(root / "api_pb2_service.cpp")
exec_clang_format(root / "api_pb2.h") exec_clang_format(root / "api_pb2.h")
exec_clang_format(root / "api_pb2.cpp") exec_clang_format(root / "api_pb2.cpp")
except ImportError: except ImportError:
pass pass
if __name__ == "__main__":
sys.exit(main())

View file

@ -61,6 +61,7 @@ solve_registry = []
def get_component_names(): def get_component_names():
# pylint: disable-next=redefined-outer-name,reimported
from esphome.loader import CORE_COMPONENTS_PATH from esphome.loader import CORE_COMPONENTS_PATH
component_names = ["esphome", "sensor", "esp32", "esp8266"] component_names = ["esphome", "sensor", "esp32", "esp8266"]
@ -82,9 +83,12 @@ def load_components():
components[domain] = get_component(domain) components[domain] = get_component(domain)
# pylint: disable=wrong-import-position
from esphome.const import CONF_TYPE, KEY_CORE from esphome.const import CONF_TYPE, KEY_CORE
from esphome.core import CORE from esphome.core import CORE
# pylint: enable=wrong-import-position
CORE.data[KEY_CORE] = {} CORE.data[KEY_CORE] = {}
load_components() load_components()
@ -114,7 +118,7 @@ def write_file(name, obj):
def delete_extra_files(keep_names): def delete_extra_files(keep_names):
for d in os.listdir(args.output_path): for d in os.listdir(args.output_path):
if d.endswith(".json") and not d[:-5] in keep_names: if d.endswith(".json") and d[:-5] not in keep_names:
os.remove(os.path.join(args.output_path, d)) os.remove(os.path.join(args.output_path, d))
print(f"Deleted {d}") print(f"Deleted {d}")
@ -552,11 +556,11 @@ def shrink():
s = f"{domain}.{schema_name}" s = f"{domain}.{schema_name}"
if ( if (
not s.endswith("." + S_CONFIG_SCHEMA) not s.endswith("." + S_CONFIG_SCHEMA)
and s not in referenced_schemas.keys() and s not in referenced_schemas
and not is_platform_schema(s) and not is_platform_schema(s)
): ):
print(f"Removing {s}") print(f"Removing {s}")
output[domain][S_SCHEMAS].pop(schema_name) domain_schemas[S_SCHEMAS].pop(schema_name)
def build_schema(): def build_schema():
@ -564,7 +568,7 @@ def build_schema():
# check esphome was not loaded globally (IDE auto imports) # check esphome was not loaded globally (IDE auto imports)
if len(ejs.extended_schemas) == 0: if len(ejs.extended_schemas) == 0:
raise Exception( raise LookupError(
"no data collected. Did you globally import an ESPHome component?" "no data collected. Did you globally import an ESPHome component?"
) )
@ -703,7 +707,7 @@ def convert(schema, config_var, path):
if schema_instance is schema: if schema_instance is schema:
assert S_CONFIG_VARS not in config_var assert S_CONFIG_VARS not in config_var
assert S_EXTENDS not in config_var assert S_EXTENDS not in config_var
if not S_TYPE in config_var: if S_TYPE not in config_var:
config_var[S_TYPE] = S_SCHEMA config_var[S_TYPE] = S_SCHEMA
# assert config_var[S_TYPE] == S_SCHEMA # assert config_var[S_TYPE] == S_SCHEMA
@ -765,9 +769,9 @@ def convert(schema, config_var, path):
elif schema == automation.validate_potentially_and_condition: elif schema == automation.validate_potentially_and_condition:
config_var[S_TYPE] = "registry" config_var[S_TYPE] = "registry"
config_var["registry"] = "condition" config_var["registry"] = "condition"
elif schema == cv.int_ or schema == cv.int_range: elif schema in (cv.int_, cv.int_range):
config_var[S_TYPE] = "integer" config_var[S_TYPE] = "integer"
elif schema == cv.string or schema == cv.string_strict or schema == cv.valid_name: elif schema in (cv.string, cv.string_strict, cv.valid_name):
config_var[S_TYPE] = "string" config_var[S_TYPE] = "string"
elif isinstance(schema, vol.Schema): elif isinstance(schema, vol.Schema):
@ -779,6 +783,7 @@ def convert(schema, config_var, path):
config_var |= pin_validators[repr_schema] config_var |= pin_validators[repr_schema]
config_var[S_TYPE] = "pin" config_var[S_TYPE] = "pin"
# pylint: disable-next=too-many-nested-blocks
elif repr_schema in ejs.hidden_schemas: elif repr_schema in ejs.hidden_schemas:
schema_type = ejs.hidden_schemas[repr_schema] schema_type = ejs.hidden_schemas[repr_schema]
@ -869,7 +874,7 @@ def convert(schema, config_var, path):
config_var["use_id_type"] = str(data.base) config_var["use_id_type"] = str(data.base)
config_var[S_TYPE] = "use_id" config_var[S_TYPE] = "use_id"
else: else:
raise Exception("Unknown extracted schema type") raise TypeError("Unknown extracted schema type")
elif config_var.get("key") == "GeneratedID": elif config_var.get("key") == "GeneratedID":
if path.startswith("i2c/CONFIG_SCHEMA/") and path.endswith("/id"): if path.startswith("i2c/CONFIG_SCHEMA/") and path.endswith("/id"):
config_var["id_type"] = { config_var["id_type"] = {
@ -884,7 +889,7 @@ def convert(schema, config_var, path):
elif path == "pins/esp32/val 1/id": elif path == "pins/esp32/val 1/id":
config_var["id_type"] = "pin" config_var["id_type"] = "pin"
else: else:
raise Exception("Cannot determine id_type for " + path) raise TypeError("Cannot determine id_type for " + path)
elif repr_schema in ejs.registry_schemas: elif repr_schema in ejs.registry_schemas:
solve_registry.append((ejs.registry_schemas[repr_schema], config_var)) solve_registry.append((ejs.registry_schemas[repr_schema], config_var))
@ -948,11 +953,7 @@ def convert_keys(converted, schema, path):
result["key"] = "GeneratedID" result["key"] = "GeneratedID"
elif isinstance(k, cv.Required): elif isinstance(k, cv.Required):
result["key"] = "Required" result["key"] = "Required"
elif ( elif isinstance(k, (cv.Optional, cv.Inclusive, cv.Exclusive)):
isinstance(k, cv.Optional)
or isinstance(k, cv.Inclusive)
or isinstance(k, cv.Exclusive)
):
result["key"] = "Optional" result["key"] = "Optional"
else: else:
converted["key"] = "String" converted["key"] = "String"

View file

@ -2,7 +2,6 @@
import argparse import argparse
import re import re
import subprocess
from dataclasses import dataclass from dataclasses import dataclass
import sys import sys
@ -40,12 +39,12 @@ class Version:
def sub(path, pattern, repl, expected_count=1): def sub(path, pattern, repl, expected_count=1):
with open(path) as fh: with open(path, encoding="utf-8") as fh:
content = fh.read() content = fh.read()
content, count = re.subn(pattern, repl, content, flags=re.MULTILINE) content, count = re.subn(pattern, repl, content, flags=re.MULTILINE)
if expected_count is not None: if expected_count is not None:
assert count == expected_count, f"Pattern {pattern} replacement failed!" assert count == expected_count, f"Pattern {pattern} replacement failed!"
with open(path, "w") as fh: with open(path, "w", encoding="utf-8") as fh:
fh.write(content) fh.write(content)

View file

@ -1,10 +1,8 @@
#!/usr/bin/env python3 #!/usr/bin/env python3
from helpers import styled, print_error_for_file, git_ls_files, filter_changed
import argparse import argparse
import codecs import codecs
import collections import collections
import colorama
import fnmatch import fnmatch
import functools import functools
import os.path import os.path
@ -12,6 +10,9 @@ import re
import sys import sys
import time import time
import colorama
from helpers import filter_changed, git_ls_files, print_error_for_file, styled
sys.path.append(os.path.dirname(__file__)) sys.path.append(os.path.dirname(__file__))
@ -30,31 +31,6 @@ def find_all(a_str, sub):
column += len(sub) column += len(sub)
colorama.init()
parser = argparse.ArgumentParser()
parser.add_argument(
"files", nargs="*", default=[], help="files to be processed (regex on path)"
)
parser.add_argument(
"-c", "--changed", action="store_true", help="Only run on changed files"
)
parser.add_argument(
"--print-slowest", action="store_true", help="Print the slowest checks"
)
args = parser.parse_args()
EXECUTABLE_BIT = git_ls_files()
files = list(EXECUTABLE_BIT.keys())
# Match against re
file_name_re = re.compile("|".join(args.files))
files = [p for p in files if file_name_re.search(p)]
if args.changed:
files = filter_changed(files)
files.sort()
file_types = ( file_types = (
".h", ".h",
".c", ".c",
@ -86,6 +62,30 @@ ignore_types = (".ico", ".png", ".woff", ".woff2", "")
LINT_FILE_CHECKS = [] LINT_FILE_CHECKS = []
LINT_CONTENT_CHECKS = [] LINT_CONTENT_CHECKS = []
LINT_POST_CHECKS = [] LINT_POST_CHECKS = []
EXECUTABLE_BIT = {}
errors = collections.defaultdict(list)
def add_errors(fname, errs):
if not isinstance(errs, list):
errs = [errs]
for err in errs:
if err is None:
continue
try:
lineno, col, msg = err
except ValueError:
lineno = 1
col = 1
msg = err
if not isinstance(msg, str):
raise ValueError("Error is not instance of string!")
if not isinstance(lineno, int):
raise ValueError("Line number is not an int!")
if not isinstance(col, int):
raise ValueError("Column number is not an int!")
errors[fname].append((lineno, col, msg))
def run_check(lint_obj, fname, *args): def run_check(lint_obj, fname, *args):
@ -155,7 +155,7 @@ def lint_re_check(regex, **kwargs):
def decorator(func): def decorator(func):
@functools.wraps(func) @functools.wraps(func)
def new_func(fname, content): def new_func(fname, content):
errors = [] errs = []
for match in prog.finditer(content): for match in prog.finditer(content):
if "NOLINT" in match.group(0): if "NOLINT" in match.group(0):
continue continue
@ -165,8 +165,8 @@ def lint_re_check(regex, **kwargs):
err = func(fname, match) err = func(fname, match)
if err is None: if err is None:
continue continue
errors.append((lineno, col + 1, err)) errs.append((lineno, col + 1, err))
return errors return errs
return decor(new_func) return decor(new_func)
@ -182,13 +182,13 @@ def lint_content_find_check(find, only_first=False, **kwargs):
find_ = find find_ = find
if callable(find): if callable(find):
find_ = find(fname, content) find_ = find(fname, content)
errors = [] errs = []
for line, col in find_all(content, find_): for line, col in find_all(content, find_):
err = func(fname) err = func(fname)
errors.append((line + 1, col + 1, err)) errs.append((line + 1, col + 1, err))
if only_first: if only_first:
break break
return errors return errs
return decor(new_func) return decor(new_func)
@ -235,8 +235,8 @@ def lint_executable_bit(fname):
ex = EXECUTABLE_BIT[fname] ex = EXECUTABLE_BIT[fname]
if ex != 100644: if ex != 100644:
return ( return (
"File has invalid executable bit {}. If running from a windows machine please " f"File has invalid executable bit {ex}. If running from a windows machine please "
"see disabling executable bit in git.".format(ex) "see disabling executable bit in git."
) )
return None return None
@ -285,8 +285,8 @@ def lint_no_defines(fname, match):
s = highlight(f"static const uint8_t {match.group(1)} = {match.group(2)};") s = highlight(f"static const uint8_t {match.group(1)} = {match.group(2)};")
return ( return (
"#define macros for integer constants are not allowed, please use " "#define macros for integer constants are not allowed, please use "
"{} style instead (replace uint8_t with the appropriate " f"{s} style instead (replace uint8_t with the appropriate "
"datatype). See also Google style guide.".format(s) "datatype). See also Google style guide."
) )
@ -296,11 +296,11 @@ def lint_no_long_delays(fname, match):
if duration_ms < 50: if duration_ms < 50:
return None return None
return ( return (
"{} - long calls to delay() are not allowed in ESPHome because everything executes " f"{highlight(match.group(0).strip())} - long calls to delay() are not allowed "
"in one thread. Calling delay() will block the main thread and slow down ESPHome.\n" "in ESPHome because everything executes in one thread. Calling delay() will "
"block the main thread and slow down ESPHome.\n"
"If there's no way to work around the delay() and it doesn't execute often, please add " "If there's no way to work around the delay() and it doesn't execute often, please add "
"a '// NOLINT' comment to the line." "a '// NOLINT' comment to the line."
"".format(highlight(match.group(0).strip()))
) )
@ -311,28 +311,28 @@ def lint_const_ordered(fname, content):
Reason: Otherwise people add it to the end, and then that results in merge conflicts. Reason: Otherwise people add it to the end, and then that results in merge conflicts.
""" """
lines = content.splitlines() lines = content.splitlines()
errors = [] errs = []
for start in ["CONF_", "ICON_", "UNIT_"]: for start in ["CONF_", "ICON_", "UNIT_"]:
matching = [ matching = [
(i + 1, line) for i, line in enumerate(lines) if line.startswith(start) (i + 1, line) for i, line in enumerate(lines) if line.startswith(start)
] ]
ordered = list(sorted(matching, key=lambda x: x[1].replace("_", " "))) ordered = list(sorted(matching, key=lambda x: x[1].replace("_", " ")))
ordered = [(mi, ol) for (mi, _), (_, ol) in zip(matching, ordered)] ordered = [(mi, ol) for (mi, _), (_, ol) in zip(matching, ordered)]
for (mi, ml), (oi, ol) in zip(matching, ordered): for (mi, mline), (_, ol) in zip(matching, ordered):
if ml == ol: if mline == ol:
continue continue
target = next(i for i, l in ordered if l == ml) target = next(i for i, line in ordered if line == mline)
target_text = next(l for i, l in matching if target == i) target_text = next(line for i, line in matching if target == i)
errors.append( errs.append(
( (
mi, mi,
1, 1,
f"Constant {highlight(ml)} is not ordered, please make sure all " f"Constant {highlight(mline)} is not ordered, please make sure all "
f"constants are ordered. See line {mi} (should go to line {target}, " f"constants are ordered. See line {mi} (should go to line {target}, "
f"{target_text})", f"{target_text})",
) )
) )
return errors return errs
@lint_re_check(r'^\s*CONF_([A-Z_0-9a-z]+)\s+=\s+[\'"](.*?)[\'"]\s*?$', include=["*.py"]) @lint_re_check(r'^\s*CONF_([A-Z_0-9a-z]+)\s+=\s+[\'"](.*?)[\'"]\s*?$', include=["*.py"])
@ -344,15 +344,14 @@ def lint_conf_matches(fname, match):
if const_norm == value_norm: if const_norm == value_norm:
return None return None
return ( return (
"Constant {} does not match value {}! Please make sure the constant's name matches its " f"Constant {highlight('CONF_' + const)} does not match value {highlight(value)}! "
"value!" "Please make sure the constant's name matches its value!"
"".format(highlight("CONF_" + const), highlight(value))
) )
CONF_RE = r'^(CONF_[a-zA-Z0-9_]+)\s*=\s*[\'"].*?[\'"]\s*?$' CONF_RE = r'^(CONF_[a-zA-Z0-9_]+)\s*=\s*[\'"].*?[\'"]\s*?$'
with codecs.open("esphome/const.py", "r", encoding="utf-8") as f_handle: with codecs.open("esphome/const.py", "r", encoding="utf-8") as const_f_handle:
constants_content = f_handle.read() constants_content = const_f_handle.read()
CONSTANTS = [m.group(1) for m in re.finditer(CONF_RE, constants_content, re.MULTILINE)] CONSTANTS = [m.group(1) for m in re.finditer(CONF_RE, constants_content, re.MULTILINE)]
CONSTANTS_USES = collections.defaultdict(list) CONSTANTS_USES = collections.defaultdict(list)
@ -365,8 +364,8 @@ def lint_conf_from_const_py(fname, match):
CONSTANTS_USES[name].append(fname) CONSTANTS_USES[name].append(fname)
return None return None
return ( return (
"Constant {} has already been defined in const.py - please import the constant from " f"Constant {highlight(name)} has already been defined in const.py - "
"const.py directly.".format(highlight(name)) "please import the constant from const.py directly."
) )
@ -473,16 +472,15 @@ def lint_no_byte_datatype(fname, match):
@lint_post_check @lint_post_check
def lint_constants_usage(): def lint_constants_usage():
errors = [] errs = []
for constant, uses in CONSTANTS_USES.items(): for constant, uses in CONSTANTS_USES.items():
if len(uses) < 4: if len(uses) < 4:
continue continue
errors.append( errs.append(
"Constant {} is defined in {} files. Please move all definitions of the " f"Constant {highlight(constant)} is defined in {len(uses)} files. Please move all definitions of the "
"constant to const.py (Uses: {})" f"constant to const.py (Uses: {', '.join(uses)})"
"".format(highlight(constant), len(uses), ", ".join(uses))
) )
return errors return errs
def relative_cpp_search_text(fname, content): def relative_cpp_search_text(fname, content):
@ -553,7 +551,7 @@ def lint_namespace(fname, content):
return ( return (
"Invalid namespace found in C++ file. All integration C++ files should put all " "Invalid namespace found in C++ file. All integration C++ files should put all "
"functions in a separate namespace that matches the integration's name. " "functions in a separate namespace that matches the integration's name. "
"Please make sure the file contains {}".format(highlight(search)) f"Please make sure the file contains {highlight(search)}"
) )
@ -639,31 +637,34 @@ def lint_log_in_header(fname):
) )
errors = collections.defaultdict(list) def main():
colorama.init()
parser = argparse.ArgumentParser()
parser.add_argument(
"files", nargs="*", default=[], help="files to be processed (regex on path)"
)
parser.add_argument(
"-c", "--changed", action="store_true", help="Only run on changed files"
)
parser.add_argument(
"--print-slowest", action="store_true", help="Print the slowest checks"
)
args = parser.parse_args()
def add_errors(fname, errs): global EXECUTABLE_BIT
if not isinstance(errs, list): EXECUTABLE_BIT = git_ls_files()
errs = [errs] files = list(EXECUTABLE_BIT.keys())
for err in errs: # Match against re
if err is None: file_name_re = re.compile("|".join(args.files))
continue files = [p for p in files if file_name_re.search(p)]
try:
lineno, col, msg = err
except ValueError:
lineno = 1
col = 1
msg = err
if not isinstance(msg, str):
raise ValueError("Error is not instance of string!")
if not isinstance(lineno, int):
raise ValueError("Line number is not an int!")
if not isinstance(col, int):
raise ValueError("Column number is not an int!")
errors[fname].append((lineno, col, msg))
if args.changed:
files = filter_changed(files)
for fname in files: files.sort()
for fname in files:
_, ext = os.path.splitext(fname) _, ext = os.path.splitext(fname)
run_checks(LINT_FILE_CHECKS, fname, fname) run_checks(LINT_FILE_CHECKS, fname, fname)
if ext in ignore_types: if ext in ignore_types:
@ -679,9 +680,9 @@ for fname in files:
continue continue
run_checks(LINT_CONTENT_CHECKS, fname, fname, content) run_checks(LINT_CONTENT_CHECKS, fname, fname, content)
run_checks(LINT_POST_CHECKS, "POST") run_checks(LINT_POST_CHECKS, "POST")
for f, errs in sorted(errors.items()): for f, errs in sorted(errors.items()):
bold = functools.partial(styled, colorama.Style.BRIGHT) bold = functools.partial(styled, colorama.Style.BRIGHT)
bold_red = functools.partial(styled, (colorama.Style.BRIGHT, colorama.Fore.RED)) bold_red = functools.partial(styled, (colorama.Style.BRIGHT, colorama.Fore.RED))
err_str = ( err_str = (
@ -690,7 +691,7 @@ for f, errs in sorted(errors.items()):
) )
print_error_for_file(f, "\n".join(err_str)) print_error_for_file(f, "\n".join(err_str))
if args.print_slowest: if args.print_slowest:
lint_times = [] lint_times = []
for lint in LINT_FILE_CHECKS + LINT_CONTENT_CHECKS + LINT_POST_CHECKS: for lint in LINT_FILE_CHECKS + LINT_CONTENT_CHECKS + LINT_POST_CHECKS:
durations = lint.get("durations", []) durations = lint.get("durations", [])
@ -701,4 +702,8 @@ if args.print_slowest:
print(f" - '{name}' took {dur:.2f}s total (ran on {invocations} files)") print(f" - '{name}' took {dur:.2f}s total (ran on {invocations} files)")
print(f"Total time measured: {sum(x[0] for x in lint_times):.2f}s") print(f"Total time measured: {sum(x[0] for x in lint_times):.2f}s")
sys.exit(len(errors)) return len(errors)
if __name__ == "__main__":
sys.exit(main())

View file

@ -1,10 +1,11 @@
import colorama import json
import os.path import os.path
import re import re
import subprocess import subprocess
import json
from pathlib import Path from pathlib import Path
import colorama
root_path = os.path.abspath(os.path.normpath(os.path.join(__file__, "..", ".."))) root_path = os.path.abspath(os.path.normpath(os.path.join(__file__, "..", "..")))
basepath = os.path.join(root_path, "esphome") basepath = os.path.join(root_path, "esphome")
temp_folder = os.path.join(root_path, ".temp") temp_folder = os.path.join(root_path, ".temp")
@ -44,7 +45,7 @@ def build_all_include():
content = "\n".join(headers) content = "\n".join(headers)
p = Path(temp_header_file) p = Path(temp_header_file)
p.parent.mkdir(exist_ok=True) p.parent.mkdir(exist_ok=True)
p.write_text(content) p.write_text(content, encoding="utf-8")
def walk_files(path): def walk_files(path):
@ -54,14 +55,14 @@ def walk_files(path):
def get_output(*args): def get_output(*args):
proc = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE) with subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as proc:
output, err = proc.communicate() output, _ = proc.communicate()
return output.decode("utf-8") return output.decode("utf-8")
def get_err(*args): def get_err(*args):
proc = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE) with subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as proc:
output, err = proc.communicate() _, err = proc.communicate()
return err.decode("utf-8") return err.decode("utf-8")
@ -78,7 +79,7 @@ def changed_files():
merge_base = splitlines_no_ends(get_output(*command))[0] merge_base = splitlines_no_ends(get_output(*command))[0]
break break
# pylint: disable=bare-except # pylint: disable=bare-except
except: except: # noqa: E722
pass pass
else: else:
raise ValueError("Git not configured") raise ValueError("Git not configured")
@ -103,7 +104,7 @@ def filter_changed(files):
def filter_grep(files, value): def filter_grep(files, value):
matched = [] matched = []
for file in files: for file in files:
with open(file) as handle: with open(file, encoding="utf-8") as handle:
contents = handle.read() contents = handle.read()
if value in contents: if value in contents:
matched.append(file) matched.append(file)
@ -114,8 +115,8 @@ def git_ls_files(patterns=None):
command = ["git", "ls-files", "-s"] command = ["git", "ls-files", "-s"]
if patterns is not None: if patterns is not None:
command.extend(patterns) command.extend(patterns)
proc = subprocess.Popen(command, stdout=subprocess.PIPE) with subprocess.Popen(command, stdout=subprocess.PIPE) as proc:
output, err = proc.communicate() output, _ = proc.communicate()
lines = [x.split() for x in output.decode("utf-8").splitlines()] lines = [x.split() for x in output.decode("utf-8").splitlines()]
return {s[3].strip(): int(s[0]) for s in lines} return {s[3].strip(): int(s[0]) for s in lines}

View file

@ -2,6 +2,7 @@
import re import re
# pylint: disable=import-error
from homeassistant.components.binary_sensor import BinarySensorDeviceClass from homeassistant.components.binary_sensor import BinarySensorDeviceClass
from homeassistant.components.button import ButtonDeviceClass from homeassistant.components.button import ButtonDeviceClass
from homeassistant.components.cover import CoverDeviceClass from homeassistant.components.cover import CoverDeviceClass
@ -9,6 +10,8 @@ from homeassistant.components.number import NumberDeviceClass
from homeassistant.components.sensor import SensorDeviceClass from homeassistant.components.sensor import SensorDeviceClass
from homeassistant.components.switch import SwitchDeviceClass from homeassistant.components.switch import SwitchDeviceClass
# pylint: enable=import-error
BLOCKLIST = ( BLOCKLIST = (
# requires special support on HA side # requires special support on HA side
"enum", "enum",
@ -25,10 +28,10 @@ DOMAINS = {
def sub(path, pattern, repl): def sub(path, pattern, repl):
with open(path) as handle: with open(path, encoding="utf-8") as handle:
content = handle.read() content = handle.read()
content = re.sub(pattern, repl, content, flags=re.MULTILINE) content = re.sub(pattern, repl, content, flags=re.MULTILINE)
with open(path, "w") as handle: with open(path, "w", encoding="utf-8") as handle:
handle.write(content) handle.write(content)