Specify src path for isort. (#8867)
This commit is contained in:
@@ -34,12 +34,12 @@ from pyspark.sql.types import (
|
||||
ShortType,
|
||||
)
|
||||
from scipy.special import expit, softmax # pylint: disable=no-name-in-module
|
||||
from xgboost.compat import is_cudf_available
|
||||
from xgboost.core import Booster
|
||||
from xgboost.training import train as worker_train
|
||||
|
||||
import xgboost
|
||||
from xgboost import XGBClassifier, XGBRanker, XGBRegressor
|
||||
from xgboost.compat import is_cudf_available
|
||||
from xgboost.core import Booster
|
||||
from xgboost.training import train as worker_train
|
||||
|
||||
from .data import (
|
||||
_read_csr_matrix_from_unwrapped_spark_vec,
|
||||
|
||||
@@ -6,9 +6,9 @@ from typing import Any, Callable, Dict, Iterator, List, Optional, Sequence, Tupl
|
||||
import numpy as np
|
||||
import pandas as pd
|
||||
from scipy.sparse import csr_matrix
|
||||
from xgboost.compat import concat
|
||||
|
||||
from xgboost import DataIter, DMatrix, QuantileDMatrix, XGBModel
|
||||
from xgboost.compat import concat
|
||||
|
||||
from .._typing import ArrayLike
|
||||
from ..core import _convert_ntree_limit
|
||||
|
||||
@@ -8,6 +8,7 @@ import uuid
|
||||
from pyspark import SparkFiles, cloudpickle
|
||||
from pyspark.ml.util import DefaultParamsReader, DefaultParamsWriter, MLReader, MLWriter
|
||||
from pyspark.sql import SparkSession
|
||||
|
||||
from xgboost.core import Booster
|
||||
|
||||
from .utils import get_class_name, get_logger
|
||||
|
||||
@@ -8,9 +8,9 @@ from typing import Any, Callable, Dict, Set, Type
|
||||
import pyspark
|
||||
from pyspark import BarrierTaskContext, SparkContext
|
||||
from pyspark.sql.session import SparkSession
|
||||
from xgboost.tracker import RabitTracker
|
||||
|
||||
from xgboost import collective
|
||||
from xgboost.tracker import RabitTracker
|
||||
|
||||
|
||||
def get_class_name(cls: Type) -> str:
|
||||
|
||||
@@ -33,10 +33,10 @@ from urllib import request
|
||||
import numpy as np
|
||||
import pytest
|
||||
from scipy import sparse
|
||||
from xgboost.core import ArrayLike
|
||||
from xgboost.sklearn import SklObjective
|
||||
|
||||
import xgboost as xgb
|
||||
from xgboost.core import ArrayLike
|
||||
from xgboost.sklearn import SklObjective
|
||||
|
||||
hypothesis = pytest.importorskip("hypothesis")
|
||||
|
||||
|
||||
@@ -2,9 +2,9 @@
|
||||
import numpy as np
|
||||
from dask import array as da
|
||||
from distributed import Client
|
||||
from xgboost.testing.updater import get_basescore
|
||||
|
||||
import xgboost as xgb
|
||||
from xgboost.testing.updater import get_basescore
|
||||
|
||||
|
||||
def check_init_estimation_clf(tree_method: str, client: Client) -> None:
|
||||
|
||||
@@ -2,6 +2,7 @@
|
||||
from typing import Any, Generator, Tuple, Union
|
||||
|
||||
import numpy as np
|
||||
|
||||
from xgboost.data import pandas_pyarrow_mapper
|
||||
|
||||
|
||||
|
||||
@@ -8,9 +8,9 @@ import tempfile
|
||||
from typing import Any, Callable, Dict, Type
|
||||
|
||||
import numpy as np
|
||||
from xgboost._typing import ArrayLike
|
||||
|
||||
import xgboost as xgb
|
||||
from xgboost._typing import ArrayLike
|
||||
|
||||
|
||||
def validate_leaf_output(leaf: np.ndarray, num_parallel_tree: int) -> None:
|
||||
|
||||
@@ -4,9 +4,9 @@ from functools import partial, update_wrapper
|
||||
from typing import Dict
|
||||
|
||||
import numpy as np
|
||||
import xgboost.testing as tm
|
||||
|
||||
import xgboost as xgb
|
||||
import xgboost.testing as tm
|
||||
|
||||
|
||||
def get_basescore(model: xgb.XGBModel) -> float:
|
||||
@@ -78,6 +78,7 @@ def check_quantile_loss(tree_method: str, weighted: bool) -> None:
|
||||
"""Test for quantile loss."""
|
||||
from sklearn.datasets import make_regression
|
||||
from sklearn.metrics import mean_pinball_loss
|
||||
|
||||
from xgboost.sklearn import _metric_decorator
|
||||
|
||||
n_samples = 4096
|
||||
|
||||
Reference in New Issue
Block a user