Specify src path for isort. (#8867)

This commit is contained in:
Jiaming Yuan
2023-03-06 17:30:27 +08:00
committed by GitHub
parent 4d665b3fb0
commit 6a892ce281
37 changed files with 59 additions and 48 deletions

View File

@@ -34,12 +34,12 @@ from pyspark.sql.types import (
ShortType,
)
from scipy.special import expit, softmax # pylint: disable=no-name-in-module
from xgboost.compat import is_cudf_available
from xgboost.core import Booster
from xgboost.training import train as worker_train
import xgboost
from xgboost import XGBClassifier, XGBRanker, XGBRegressor
from xgboost.compat import is_cudf_available
from xgboost.core import Booster
from xgboost.training import train as worker_train
from .data import (
_read_csr_matrix_from_unwrapped_spark_vec,

View File

@@ -6,9 +6,9 @@ from typing import Any, Callable, Dict, Iterator, List, Optional, Sequence, Tupl
import numpy as np
import pandas as pd
from scipy.sparse import csr_matrix
from xgboost.compat import concat
from xgboost import DataIter, DMatrix, QuantileDMatrix, XGBModel
from xgboost.compat import concat
from .._typing import ArrayLike
from ..core import _convert_ntree_limit

View File

@@ -8,6 +8,7 @@ import uuid
from pyspark import SparkFiles, cloudpickle
from pyspark.ml.util import DefaultParamsReader, DefaultParamsWriter, MLReader, MLWriter
from pyspark.sql import SparkSession
from xgboost.core import Booster
from .utils import get_class_name, get_logger

View File

@@ -8,9 +8,9 @@ from typing import Any, Callable, Dict, Set, Type
import pyspark
from pyspark import BarrierTaskContext, SparkContext
from pyspark.sql.session import SparkSession
from xgboost.tracker import RabitTracker
from xgboost import collective
from xgboost.tracker import RabitTracker
def get_class_name(cls: Type) -> str:

View File

@@ -33,10 +33,10 @@ from urllib import request
import numpy as np
import pytest
from scipy import sparse
from xgboost.core import ArrayLike
from xgboost.sklearn import SklObjective
import xgboost as xgb
from xgboost.core import ArrayLike
from xgboost.sklearn import SklObjective
hypothesis = pytest.importorskip("hypothesis")

View File

@@ -2,9 +2,9 @@
import numpy as np
from dask import array as da
from distributed import Client
from xgboost.testing.updater import get_basescore
import xgboost as xgb
from xgboost.testing.updater import get_basescore
def check_init_estimation_clf(tree_method: str, client: Client) -> None:

View File

@@ -2,6 +2,7 @@
from typing import Any, Generator, Tuple, Union
import numpy as np
from xgboost.data import pandas_pyarrow_mapper

View File

@@ -8,9 +8,9 @@ import tempfile
from typing import Any, Callable, Dict, Type
import numpy as np
from xgboost._typing import ArrayLike
import xgboost as xgb
from xgboost._typing import ArrayLike
def validate_leaf_output(leaf: np.ndarray, num_parallel_tree: int) -> None:

View File

@@ -4,9 +4,9 @@ from functools import partial, update_wrapper
from typing import Dict
import numpy as np
import xgboost.testing as tm
import xgboost as xgb
import xgboost.testing as tm
def get_basescore(model: xgb.XGBModel) -> float:
@@ -78,6 +78,7 @@ def check_quantile_loss(tree_method: str, weighted: bool) -> None:
"""Test for quantile loss."""
from sklearn.datasets import make_regression
from sklearn.metrics import mean_pinball_loss
from xgboost.sklearn import _metric_decorator
n_samples = 4096